var/home/core/zuul-output/0000755000175000017500000000000015073370253014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073411502015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000007104442215073411473017706 0ustar rootrootOct 14 06:49:42 crc systemd[1]: Starting Kubernetes Kubelet... Oct 14 06:49:42 crc restorecon[4796]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:42 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 06:49:43 crc restorecon[4796]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 14 06:49:44 crc kubenswrapper[5018]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.192033 5018 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217105 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217159 5018 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217172 5018 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217183 5018 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217193 5018 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217203 5018 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217213 5018 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217221 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217230 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217240 5018 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217250 5018 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217260 5018 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217268 5018 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217277 5018 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217285 5018 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217293 5018 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217301 5018 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217311 5018 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217319 5018 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217331 5018 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217344 5018 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217353 5018 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217364 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217374 5018 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217384 5018 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217394 5018 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217417 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217427 5018 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217436 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217445 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217454 5018 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217463 5018 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217472 5018 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217480 5018 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217489 5018 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217500 5018 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217511 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217520 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217531 5018 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217540 5018 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217548 5018 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217557 5018 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217565 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217576 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217584 5018 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217593 5018 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217601 5018 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217610 5018 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217627 5018 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217637 5018 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217672 5018 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217682 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217691 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217700 5018 feature_gate.go:330] unrecognized feature gate: Example Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217708 5018 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217717 5018 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217725 5018 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217734 5018 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217743 5018 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217752 5018 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217760 5018 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217770 5018 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217779 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217788 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217798 5018 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217807 5018 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217816 5018 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217824 5018 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217835 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217844 5018 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.217853 5018 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.218025 5018 flags.go:64] FLAG: --address="0.0.0.0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.218044 5018 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222587 5018 flags.go:64] FLAG: --anonymous-auth="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222604 5018 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222646 5018 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222657 5018 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222669 5018 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222682 5018 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222693 5018 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222704 5018 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222715 5018 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222730 5018 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222741 5018 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222751 5018 flags.go:64] FLAG: --cgroup-root="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222761 5018 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222771 5018 flags.go:64] FLAG: --client-ca-file="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222781 5018 flags.go:64] FLAG: --cloud-config="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222792 5018 flags.go:64] FLAG: --cloud-provider="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222803 5018 flags.go:64] FLAG: --cluster-dns="[]" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222817 5018 flags.go:64] FLAG: --cluster-domain="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222827 5018 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222837 5018 flags.go:64] FLAG: --config-dir="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222847 5018 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222857 5018 flags.go:64] FLAG: --container-log-max-files="5" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222870 5018 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222880 5018 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222890 5018 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222900 5018 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222909 5018 flags.go:64] FLAG: --contention-profiling="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222920 5018 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222929 5018 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222940 5018 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222949 5018 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222962 5018 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222972 5018 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222982 5018 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.222991 5018 flags.go:64] FLAG: --enable-load-reader="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223001 5018 flags.go:64] FLAG: --enable-server="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223012 5018 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223025 5018 flags.go:64] FLAG: --event-burst="100" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223037 5018 flags.go:64] FLAG: --event-qps="50" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223048 5018 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223059 5018 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223069 5018 flags.go:64] FLAG: --eviction-hard="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223081 5018 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223091 5018 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223101 5018 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223113 5018 flags.go:64] FLAG: --eviction-soft="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223123 5018 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223133 5018 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223143 5018 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223153 5018 flags.go:64] FLAG: --experimental-mounter-path="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223163 5018 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223173 5018 flags.go:64] FLAG: --fail-swap-on="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223182 5018 flags.go:64] FLAG: --feature-gates="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223194 5018 flags.go:64] FLAG: --file-check-frequency="20s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223204 5018 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223214 5018 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223224 5018 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223234 5018 flags.go:64] FLAG: --healthz-port="10248" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223244 5018 flags.go:64] FLAG: --help="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223254 5018 flags.go:64] FLAG: --hostname-override="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223264 5018 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223274 5018 flags.go:64] FLAG: --http-check-frequency="20s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223283 5018 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223293 5018 flags.go:64] FLAG: --image-credential-provider-config="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223303 5018 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223313 5018 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223323 5018 flags.go:64] FLAG: --image-service-endpoint="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223332 5018 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223342 5018 flags.go:64] FLAG: --kube-api-burst="100" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223352 5018 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223363 5018 flags.go:64] FLAG: --kube-api-qps="50" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223373 5018 flags.go:64] FLAG: --kube-reserved="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223385 5018 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223395 5018 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223405 5018 flags.go:64] FLAG: --kubelet-cgroups="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223414 5018 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223424 5018 flags.go:64] FLAG: --lock-file="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223434 5018 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223443 5018 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223454 5018 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223467 5018 flags.go:64] FLAG: --log-json-split-stream="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223478 5018 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223488 5018 flags.go:64] FLAG: --log-text-split-stream="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223497 5018 flags.go:64] FLAG: --logging-format="text" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223507 5018 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223518 5018 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223527 5018 flags.go:64] FLAG: --manifest-url="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223537 5018 flags.go:64] FLAG: --manifest-url-header="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223550 5018 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223559 5018 flags.go:64] FLAG: --max-open-files="1000000" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223572 5018 flags.go:64] FLAG: --max-pods="110" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223581 5018 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223591 5018 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223601 5018 flags.go:64] FLAG: --memory-manager-policy="None" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223611 5018 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223627 5018 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223637 5018 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223647 5018 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223702 5018 flags.go:64] FLAG: --node-status-max-images="50" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223711 5018 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223722 5018 flags.go:64] FLAG: --oom-score-adj="-999" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223732 5018 flags.go:64] FLAG: --pod-cidr="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223741 5018 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223755 5018 flags.go:64] FLAG: --pod-manifest-path="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223764 5018 flags.go:64] FLAG: --pod-max-pids="-1" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223775 5018 flags.go:64] FLAG: --pods-per-core="0" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223785 5018 flags.go:64] FLAG: --port="10250" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223795 5018 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223805 5018 flags.go:64] FLAG: --provider-id="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223815 5018 flags.go:64] FLAG: --qos-reserved="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223825 5018 flags.go:64] FLAG: --read-only-port="10255" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223835 5018 flags.go:64] FLAG: --register-node="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223845 5018 flags.go:64] FLAG: --register-schedulable="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223855 5018 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223871 5018 flags.go:64] FLAG: --registry-burst="10" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223882 5018 flags.go:64] FLAG: --registry-qps="5" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223891 5018 flags.go:64] FLAG: --reserved-cpus="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223902 5018 flags.go:64] FLAG: --reserved-memory="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223928 5018 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223937 5018 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223948 5018 flags.go:64] FLAG: --rotate-certificates="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223958 5018 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223968 5018 flags.go:64] FLAG: --runonce="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223978 5018 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223989 5018 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.223999 5018 flags.go:64] FLAG: --seccomp-default="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224009 5018 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224019 5018 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224029 5018 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224039 5018 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224049 5018 flags.go:64] FLAG: --storage-driver-password="root" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224059 5018 flags.go:64] FLAG: --storage-driver-secure="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224069 5018 flags.go:64] FLAG: --storage-driver-table="stats" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224078 5018 flags.go:64] FLAG: --storage-driver-user="root" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224087 5018 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224099 5018 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224110 5018 flags.go:64] FLAG: --system-cgroups="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224120 5018 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224135 5018 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224145 5018 flags.go:64] FLAG: --tls-cert-file="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224154 5018 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224167 5018 flags.go:64] FLAG: --tls-min-version="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224177 5018 flags.go:64] FLAG: --tls-private-key-file="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224186 5018 flags.go:64] FLAG: --topology-manager-policy="none" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224195 5018 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224205 5018 flags.go:64] FLAG: --topology-manager-scope="container" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224215 5018 flags.go:64] FLAG: --v="2" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224232 5018 flags.go:64] FLAG: --version="false" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224244 5018 flags.go:64] FLAG: --vmodule="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224288 5018 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.224299 5018 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224510 5018 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224522 5018 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224533 5018 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224542 5018 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224551 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224559 5018 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224568 5018 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224576 5018 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224585 5018 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224597 5018 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224608 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224618 5018 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224627 5018 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224641 5018 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224673 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224681 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224690 5018 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224700 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224709 5018 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224717 5018 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224726 5018 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224734 5018 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224743 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224751 5018 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224761 5018 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224769 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224778 5018 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224786 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224795 5018 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224806 5018 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224817 5018 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224826 5018 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224835 5018 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224844 5018 feature_gate.go:330] unrecognized feature gate: Example Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224853 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224862 5018 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224870 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224878 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224889 5018 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224898 5018 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224906 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224914 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224923 5018 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224932 5018 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224940 5018 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224948 5018 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224958 5018 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224966 5018 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224976 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224986 5018 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.224994 5018 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225002 5018 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225011 5018 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225020 5018 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225028 5018 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225036 5018 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225045 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225053 5018 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225062 5018 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225070 5018 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225079 5018 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225090 5018 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225100 5018 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225110 5018 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225120 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225130 5018 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225138 5018 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225147 5018 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225156 5018 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225168 5018 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.225178 5018 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.225193 5018 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.237286 5018 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.237344 5018 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237469 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237493 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237504 5018 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237514 5018 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237523 5018 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237532 5018 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237540 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237549 5018 feature_gate.go:330] unrecognized feature gate: Example Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237557 5018 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237566 5018 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237575 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237585 5018 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237594 5018 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237604 5018 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237612 5018 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237625 5018 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237634 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237642 5018 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237672 5018 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237680 5018 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237689 5018 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237697 5018 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237705 5018 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237713 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237721 5018 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237729 5018 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237738 5018 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237745 5018 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237753 5018 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237761 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237769 5018 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237777 5018 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237790 5018 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237801 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237815 5018 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237827 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237836 5018 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237844 5018 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237853 5018 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237861 5018 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237869 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237877 5018 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237884 5018 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237893 5018 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237901 5018 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237909 5018 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237920 5018 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237930 5018 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237939 5018 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237949 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237958 5018 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237966 5018 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237975 5018 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237983 5018 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.237992 5018 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238001 5018 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238010 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238018 5018 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238026 5018 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238034 5018 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238042 5018 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238053 5018 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238063 5018 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238074 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238082 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238092 5018 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238101 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238109 5018 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238117 5018 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238125 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238135 5018 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.238149 5018 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238395 5018 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238409 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238418 5018 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238426 5018 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238434 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238442 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238452 5018 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238464 5018 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238475 5018 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238485 5018 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238494 5018 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238502 5018 feature_gate.go:330] unrecognized feature gate: Example Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238511 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238519 5018 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238527 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238535 5018 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238542 5018 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238550 5018 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238558 5018 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238566 5018 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238573 5018 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238581 5018 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238590 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238599 5018 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238607 5018 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238615 5018 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238628 5018 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238636 5018 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238644 5018 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238677 5018 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238685 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238693 5018 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238701 5018 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238709 5018 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238719 5018 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238726 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238735 5018 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238745 5018 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238754 5018 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238764 5018 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238773 5018 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238781 5018 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238789 5018 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238798 5018 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238806 5018 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238814 5018 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238822 5018 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238830 5018 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238838 5018 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238846 5018 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238854 5018 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238862 5018 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238870 5018 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238878 5018 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238887 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238896 5018 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238904 5018 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238912 5018 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238919 5018 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238927 5018 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238935 5018 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238945 5018 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238956 5018 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238965 5018 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238975 5018 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238984 5018 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.238992 5018 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.239000 5018 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.239008 5018 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.239016 5018 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.239028 5018 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.239040 5018 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.239297 5018 server.go:940] "Client rotation is on, will bootstrap in background" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.249014 5018 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.249208 5018 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.253016 5018 server.go:997] "Starting client certificate rotation" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.253067 5018 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.253339 5018 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-15 04:38:30.36399558 +0000 UTC Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.253453 5018 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2229h48m46.110548105s for next certificate rotation Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.327837 5018 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.334136 5018 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.363776 5018 log.go:25] "Validated CRI v1 runtime API" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.460263 5018 log.go:25] "Validated CRI v1 image API" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.463955 5018 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.472353 5018 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-14-06-40-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.472411 5018 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.504260 5018 manager.go:217] Machine: {Timestamp:2025-10-14 06:49:44.501047844 +0000 UTC m=+1.085094531 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ca6145eb-67b2-4317-a65f-9b8f914c2ca3 BootID:772668f9-2246-4743-84c4-813da958e906 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:30:57:2c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:30:57:2c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:95:04:a0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1d:0a:dd Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:11:5d:9e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:75:23:cb Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:85:7c:e4 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:68:23:90 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:00:69:1f:45:ad Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0a:8f:82:d2:e3:02 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.504695 5018 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.504905 5018 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.505473 5018 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.505831 5018 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.505898 5018 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.506233 5018 topology_manager.go:138] "Creating topology manager with none policy" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.506252 5018 container_manager_linux.go:303] "Creating device plugin manager" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.506789 5018 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.506841 5018 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.507070 5018 state_mem.go:36] "Initialized new in-memory state store" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.507224 5018 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.511505 5018 kubelet.go:418] "Attempting to sync node with API server" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.511548 5018 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.511576 5018 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.511598 5018 kubelet.go:324] "Adding apiserver pod source" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.511622 5018 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.516204 5018 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.518506 5018 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.519584 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.519719 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.519739 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.519873 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.521359 5018 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523127 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523167 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523183 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523197 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523223 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523236 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523250 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523274 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523289 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523304 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523344 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.523359 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.524720 5018 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.525266 5018 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.525465 5018 server.go:1280] "Started kubelet" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.527448 5018 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.527443 5018 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 14 06:49:44 crc systemd[1]: Started Kubernetes Kubelet. Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.528872 5018 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.531408 5018 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.531680 5018 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 01:50:13.868314335 +0000 UTC Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.531861 5018 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2299h0m29.336463204s for next certificate rotation Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.532064 5018 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.534699 5018 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.534753 5018 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.534983 5018 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.535443 5018 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.542961 5018 factory.go:55] Registering systemd factory Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.543021 5018 factory.go:221] Registration of the systemd container factory successfully Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.543587 5018 server.go:460] "Adding debug handlers to kubelet server" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.549680 5018 factory.go:153] Registering CRI-O factory Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.549725 5018 factory.go:221] Registration of the crio container factory successfully Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.549727 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="200ms" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.549836 5018 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.549915 5018 factory.go:103] Registering Raw factory Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.549946 5018 manager.go:1196] Started watching for new ooms in manager Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.550002 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.550132 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.550971 5018 manager.go:319] Starting recovery of all containers Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.550128 5018 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.159:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e48d51e8088a3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-14 06:49:44.525416611 +0000 UTC m=+1.109463268,LastTimestamp:2025-10-14 06:49:44.525416611 +0000 UTC m=+1.109463268,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560737 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560827 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560856 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560881 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560908 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560935 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560960 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.560986 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561014 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561039 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561063 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561092 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561119 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561148 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561172 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561198 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561224 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561251 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561277 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561304 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561327 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561353 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561377 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561407 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561432 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561459 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561559 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561593 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561628 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561684 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561807 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561837 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561869 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561895 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561932 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561957 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.561982 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562005 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562032 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562058 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562095 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562122 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562148 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562172 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562198 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562230 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562258 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562286 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562312 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562337 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562365 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562391 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562425 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562453 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562481 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562510 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562536 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562563 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562587 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562613 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562683 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562715 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562741 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562772 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562798 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562822 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562922 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562949 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.562974 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563000 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563027 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563102 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563131 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563158 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563184 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563208 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563233 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563262 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563320 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563378 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563407 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563433 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563460 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563545 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563577 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563601 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563627 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563694 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563721 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563745 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563764 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563784 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563804 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563837 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563856 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563876 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563895 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563915 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563935 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563956 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563977 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.563998 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.564018 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.564037 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.564065 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.564087 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.564109 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567591 5018 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567683 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567720 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567771 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567791 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567818 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567839 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567858 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567877 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567894 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567911 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567925 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567941 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567957 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567972 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.567994 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568010 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568026 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568044 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568058 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568083 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568099 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568115 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568129 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568143 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568158 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568184 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568202 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568218 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568235 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568249 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568263 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568278 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568294 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568316 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568333 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568349 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568364 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568378 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568391 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568403 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568416 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568430 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568443 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568454 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568470 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568484 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568497 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568508 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568525 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568539 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568552 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568564 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568577 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568591 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568604 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568621 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568634 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568661 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568673 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568686 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568699 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568713 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568725 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568739 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568754 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568766 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568777 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568790 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568803 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568814 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568828 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568841 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568853 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568865 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568876 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568887 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568898 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568908 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568921 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568932 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568944 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568958 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568970 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568983 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.568995 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569007 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569019 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569032 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569046 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569058 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569069 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569080 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569097 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569108 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569118 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569130 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569140 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569153 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569163 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569173 5018 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569183 5018 reconstruct.go:97] "Volume reconstruction finished" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.569191 5018 reconciler.go:26] "Reconciler: start to sync state" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.582627 5018 manager.go:324] Recovery completed Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.599240 5018 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.602381 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.603610 5018 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.603662 5018 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.603694 5018 kubelet.go:2335] "Starting kubelet main sync loop" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.603791 5018 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.605127 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.605198 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.605220 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: W1014 06:49:44.605486 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.605590 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.607560 5018 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.607585 5018 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.607620 5018 state_mem.go:36] "Initialized new in-memory state store" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.624907 5018 policy_none.go:49] "None policy: Start" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.626444 5018 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.626494 5018 state_mem.go:35] "Initializing new in-memory state store" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.635717 5018 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.693867 5018 manager.go:334] "Starting Device Plugin manager" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.693991 5018 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.694013 5018 server.go:79] "Starting device plugin registration server" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.694613 5018 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.694678 5018 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.694963 5018 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.695088 5018 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.695109 5018 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.704562 5018 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.704722 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.705873 5018 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.706543 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.706601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.706623 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.706873 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.707199 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.707258 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708402 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708455 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708471 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708458 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708539 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708705 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708781 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.708829 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.709760 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.709796 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.709809 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.709959 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710069 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710094 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710117 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710162 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710809 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710843 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710855 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.710981 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711112 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711159 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711276 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711328 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711873 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711903 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.711919 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.712104 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.712138 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.712946 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.712979 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.712997 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.713003 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.713021 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.713031 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.750449 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="400ms" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771481 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771537 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771572 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771599 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771656 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771690 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771720 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771786 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771849 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771889 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771922 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771960 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.771991 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.772023 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.772053 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.795187 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.796237 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.796280 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.796304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.796340 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.796932 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873187 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873236 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873257 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873275 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873320 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873348 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873449 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873517 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873569 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873451 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873679 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873701 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873718 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873763 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873796 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873805 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873824 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873767 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873855 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873882 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873885 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873925 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873932 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.873970 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.874013 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.874054 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.874091 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.874193 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.997368 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.998622 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.998674 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.998686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:44 crc kubenswrapper[5018]: I1014 06:49:44.998710 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:44 crc kubenswrapper[5018]: E1014 06:49:44.999081 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.048679 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.067624 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.091595 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.103202 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.108796 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.113191 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-028564aac1666db3f31c5b5fd304977f3759345176d0fc1415a899bf5c4f8276 WatchSource:0}: Error finding container 028564aac1666db3f31c5b5fd304977f3759345176d0fc1415a899bf5c4f8276: Status 404 returned error can't find the container with id 028564aac1666db3f31c5b5fd304977f3759345176d0fc1415a899bf5c4f8276 Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.115305 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-857c33dcb99e2f9922745d38d7736cfc91c6023f238c0ae27678c6a3e3f127ae WatchSource:0}: Error finding container 857c33dcb99e2f9922745d38d7736cfc91c6023f238c0ae27678c6a3e3f127ae: Status 404 returned error can't find the container with id 857c33dcb99e2f9922745d38d7736cfc91c6023f238c0ae27678c6a3e3f127ae Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.131337 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a904dc8b083f644785fd9d2093f0b6d34736219aa2450130ae908cf17d9725a3 WatchSource:0}: Error finding container a904dc8b083f644785fd9d2093f0b6d34736219aa2450130ae908cf17d9725a3: Status 404 returned error can't find the container with id a904dc8b083f644785fd9d2093f0b6d34736219aa2450130ae908cf17d9725a3 Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.134572 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a86d8cde26e50238d1e7fe893c9538331b845a4ed7a87cccf6c568bcea18280f WatchSource:0}: Error finding container a86d8cde26e50238d1e7fe893c9538331b845a4ed7a87cccf6c568bcea18280f: Status 404 returned error can't find the container with id a86d8cde26e50238d1e7fe893c9538331b845a4ed7a87cccf6c568bcea18280f Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.137332 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9f1044617e218143ddcf4fe5bc35ac66f7ead23789f5e69fa01cd1042d5a4b9e WatchSource:0}: Error finding container 9f1044617e218143ddcf4fe5bc35ac66f7ead23789f5e69fa01cd1042d5a4b9e: Status 404 returned error can't find the container with id 9f1044617e218143ddcf4fe5bc35ac66f7ead23789f5e69fa01cd1042d5a4b9e Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.152149 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="800ms" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.399561 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.401680 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.401754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.401780 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.401826 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.402425 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.468602 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.468745 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.526828 5018 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.608868 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a86d8cde26e50238d1e7fe893c9538331b845a4ed7a87cccf6c568bcea18280f"} Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.610582 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9f1044617e218143ddcf4fe5bc35ac66f7ead23789f5e69fa01cd1042d5a4b9e"} Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.612437 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a904dc8b083f644785fd9d2093f0b6d34736219aa2450130ae908cf17d9725a3"} Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.613399 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"857c33dcb99e2f9922745d38d7736cfc91c6023f238c0ae27678c6a3e3f127ae"} Oct 14 06:49:45 crc kubenswrapper[5018]: I1014 06:49:45.614362 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"028564aac1666db3f31c5b5fd304977f3759345176d0fc1415a899bf5c4f8276"} Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.903142 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.903472 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.953957 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="1.6s" Oct 14 06:49:45 crc kubenswrapper[5018]: W1014 06:49:45.984926 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:45 crc kubenswrapper[5018]: E1014 06:49:45.985022 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:46 crc kubenswrapper[5018]: W1014 06:49:46.078826 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:46 crc kubenswrapper[5018]: E1014 06:49:46.078997 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.203257 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.205201 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.205272 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.205298 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.205348 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:46 crc kubenswrapper[5018]: E1014 06:49:46.206110 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.526960 5018 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.620382 5018 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="81b96a52ecef8940d16ac6832ebee342ad30ed212694c4330e95e2b98c608d4d" exitCode=0 Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.620482 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"81b96a52ecef8940d16ac6832ebee342ad30ed212694c4330e95e2b98c608d4d"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.620563 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.621898 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.621957 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.621981 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.624896 5018 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="0fb471214c903c143f9af9a667b6cd0c57d90249112375b971df7ff27452c421" exitCode=0 Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.625066 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.625061 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"0fb471214c903c143f9af9a667b6cd0c57d90249112375b971df7ff27452c421"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.628580 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.628704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.628735 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.631731 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.631807 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.631842 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.635274 5018 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5" exitCode=0 Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.635461 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.636075 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.636823 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.636875 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.636894 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.640365 5018 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2" exitCode=0 Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.640410 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2"} Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.640522 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.641600 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.641677 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.641697 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.646314 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.647567 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.647689 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:46 crc kubenswrapper[5018]: I1014 06:49:46.647705 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: W1014 06:49:47.204461 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:47 crc kubenswrapper[5018]: E1014 06:49:47.204532 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.526484 5018 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:47 crc kubenswrapper[5018]: E1014 06:49:47.554787 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.159:6443: connect: connection refused" interval="3.2s" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.647667 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.647736 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.647749 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.647777 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.650535 5018 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18b8ec95cc2a4497435a46e1c3d284fa46e7fe2b089c844faeff4f82ea51210f" exitCode=0 Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.650609 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18b8ec95cc2a4497435a46e1c3d284fa46e7fe2b089c844faeff4f82ea51210f"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.650716 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.652443 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.652481 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.652498 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.653833 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.654323 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7a91b6a97324554e96b23f53977bc4653d2624df3711e122a584950fa2b463c7"} Oct 14 06:49:47 crc kubenswrapper[5018]: W1014 06:49:47.654929 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:47 crc kubenswrapper[5018]: E1014 06:49:47.655008 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.655942 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.655977 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.655990 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.661583 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.661636 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.662462 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.662491 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.662505 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.664734 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.664788 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.664802 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13"} Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.664976 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.666454 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.666487 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.666501 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.806240 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.807593 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.807655 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.807669 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:47 crc kubenswrapper[5018]: I1014 06:49:47.807694 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:47 crc kubenswrapper[5018]: E1014 06:49:47.808467 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.159:6443: connect: connection refused" node="crc" Oct 14 06:49:48 crc kubenswrapper[5018]: W1014 06:49:48.127015 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.159:6443: connect: connection refused Oct 14 06:49:48 crc kubenswrapper[5018]: E1014 06:49:48.127150 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.159:6443: connect: connection refused" logger="UnhandledError" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.673206 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100"} Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.673415 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.674825 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.674893 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.674917 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676672 5018 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="98b874047871237937a79111aa40e99647fd89e46856493ada162e63edf132f5" exitCode=0 Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676736 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676816 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676860 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676869 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.676896 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677064 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"98b874047871237937a79111aa40e99647fd89e46856493ada162e63edf132f5"} Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677820 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677832 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677906 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677928 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.677939 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.678353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.678462 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.678558 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.678691 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.679096 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:48 crc kubenswrapper[5018]: I1014 06:49:48.679123 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.688873 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"090b871e2becd8d72be94cb5794712c5f495b244487d1b97425f0a6ee6935d79"} Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.688960 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ad98701cf6764de9cb4416c34d3e01b9be9d7fc03d657cb82cbd04b32662615a"} Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.688982 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9edeccbe0539d0c63f6bc7bb102f46aa6e49c650d5bcf2944c668ea176730235"} Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.689004 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4bc1ca5b193c226644e0d7c9c30787d26b5afeecd2ecf87e41294a4fef702381"} Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.689022 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.689122 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.690556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.690615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:49 crc kubenswrapper[5018]: I1014 06:49:49.690685 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.381039 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.452749 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.453013 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.454593 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.454713 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.454732 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.698303 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3105f87790fa4ed5c1c6ded2170eef9337f5ea23b61150c1eb24ffb3d2393149"} Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.698427 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.698427 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700008 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700070 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700013 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700127 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700151 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.700094 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.934796 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.935008 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.937166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.937249 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.937265 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:50 crc kubenswrapper[5018]: I1014 06:49:50.943599 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.009042 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.011578 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.011716 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.011742 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.011782 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.367883 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.368120 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.369761 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.369817 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.369838 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.700896 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.700926 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.700947 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703043 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703103 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703122 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703851 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703890 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703908 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703918 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.703982 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:51 crc kubenswrapper[5018]: I1014 06:49:51.704007 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:52 crc kubenswrapper[5018]: I1014 06:49:52.234264 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:49:52 crc kubenswrapper[5018]: I1014 06:49:52.704419 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:52 crc kubenswrapper[5018]: I1014 06:49:52.706242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:52 crc kubenswrapper[5018]: I1014 06:49:52.706291 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:52 crc kubenswrapper[5018]: I1014 06:49:52.706311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:53 crc kubenswrapper[5018]: I1014 06:49:53.748477 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 14 06:49:53 crc kubenswrapper[5018]: I1014 06:49:53.748832 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:53 crc kubenswrapper[5018]: I1014 06:49:53.750605 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:53 crc kubenswrapper[5018]: I1014 06:49:53.750783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:53 crc kubenswrapper[5018]: I1014 06:49:53.750817 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:54 crc kubenswrapper[5018]: I1014 06:49:54.051266 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:54 crc kubenswrapper[5018]: I1014 06:49:54.051545 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:54 crc kubenswrapper[5018]: I1014 06:49:54.053183 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:54 crc kubenswrapper[5018]: I1014 06:49:54.053249 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:54 crc kubenswrapper[5018]: I1014 06:49:54.053269 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:54 crc kubenswrapper[5018]: E1014 06:49:54.706020 5018 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.367695 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.368028 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.370520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.370682 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.370710 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.375196 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.718162 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.720045 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.720095 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:49:56 crc kubenswrapper[5018]: I1014 06:49:56.720112 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:49:57 crc kubenswrapper[5018]: I1014 06:49:57.052120 5018 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 06:49:57 crc kubenswrapper[5018]: I1014 06:49:57.052283 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 06:49:58 crc kubenswrapper[5018]: W1014 06:49:58.348955 5018 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 14 06:49:58 crc kubenswrapper[5018]: I1014 06:49:58.349147 5018 trace.go:236] Trace[1230872949]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 06:49:48.346) (total time: 10002ms): Oct 14 06:49:58 crc kubenswrapper[5018]: Trace[1230872949]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (06:49:58.348) Oct 14 06:49:58 crc kubenswrapper[5018]: Trace[1230872949]: [10.002287584s] [10.002287584s] END Oct 14 06:49:58 crc kubenswrapper[5018]: E1014 06:49:58.349192 5018 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 14 06:49:58 crc kubenswrapper[5018]: I1014 06:49:58.440168 5018 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 06:49:58 crc kubenswrapper[5018]: I1014 06:49:58.440304 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 06:49:58 crc kubenswrapper[5018]: I1014 06:49:58.445636 5018 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 06:49:58 crc kubenswrapper[5018]: I1014 06:49:58.445693 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 06:49:59 crc kubenswrapper[5018]: I1014 06:49:59.997996 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 14 06:49:59 crc kubenswrapper[5018]: I1014 06:49:59.998318 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.000085 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.000133 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.000145 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.051396 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.390343 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.390740 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.392309 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.392353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.392365 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.398073 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.727656 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.728043 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.727808 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730148 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730221 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730412 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730456 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.730478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:00 crc kubenswrapper[5018]: I1014 06:50:00.749768 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 14 06:50:01 crc kubenswrapper[5018]: I1014 06:50:01.730514 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:01 crc kubenswrapper[5018]: I1014 06:50:01.732021 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:01 crc kubenswrapper[5018]: I1014 06:50:01.732077 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:01 crc kubenswrapper[5018]: I1014 06:50:01.732102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:02 crc kubenswrapper[5018]: I1014 06:50:02.772928 5018 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.412029 5018 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.416462 5018 trace.go:236] Trace[1681286668]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 06:49:51.190) (total time: 12225ms): Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[1681286668]: ---"Objects listed" error: 12225ms (06:50:03.416) Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[1681286668]: [12.225673872s] [12.225673872s] END Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.416505 5018 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.418222 5018 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.418467 5018 trace.go:236] Trace[571450919]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 06:49:52.860) (total time: 10557ms): Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[571450919]: ---"Objects listed" error: 10557ms (06:50:03.418) Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[571450919]: [10.55741276s] [10.55741276s] END Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.418517 5018 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.421128 5018 trace.go:236] Trace[167296310]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 06:49:52.296) (total time: 11124ms): Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[167296310]: ---"Objects listed" error: 11124ms (06:50:03.420) Oct 14 06:50:03 crc kubenswrapper[5018]: Trace[167296310]: [11.124932393s] [11.124932393s] END Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.421172 5018 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.432325 5018 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.474099 5018 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.474166 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.474563 5018 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.474609 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.525472 5018 apiserver.go:52] "Watching apiserver" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.530772 5018 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.531046 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.531418 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.531535 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.531612 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.531804 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.531872 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.531913 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.532049 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.532104 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.532152 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.536315 5018 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.536507 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.536685 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.536926 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.537048 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.536935 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.537277 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.538379 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.538548 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.543297 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.565532 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.576647 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.586088 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.599170 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.611188 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.620496 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.630121 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.632849 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.632903 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.632931 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.632953 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.632977 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633374 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633402 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633407 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633453 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633002 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633521 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633543 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633582 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633611 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633717 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633741 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633756 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633799 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633830 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633852 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633903 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633926 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633947 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.633968 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634006 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634049 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634025 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634071 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634355 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634407 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634418 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634852 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634929 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634992 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635057 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635164 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635202 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635238 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635277 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635313 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635350 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635384 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635421 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635459 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635498 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635534 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635575 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635695 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635740 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635783 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635822 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635857 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635891 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635935 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635970 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636009 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636054 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636106 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636151 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634511 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634647 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634859 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636200 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634873 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636260 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636293 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636329 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636366 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636408 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636489 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636537 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.637874 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.634945 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635114 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635127 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635145 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635269 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635308 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635406 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635541 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635696 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635721 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635808 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.635867 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636048 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636101 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636168 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636368 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636385 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636390 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636505 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.636655 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.637824 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.638060 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.638302 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639243 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639372 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639384 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639583 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639796 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.639996 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640001 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640098 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640104 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640166 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640239 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640359 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640528 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640720 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640779 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.640785 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641260 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641411 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641486 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641552 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641859 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641865 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641874 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641907 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641935 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641959 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.641983 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642005 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642030 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642233 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642382 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642661 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642789 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.642847 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.643421 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.643490 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.643933 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.643942 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644011 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644092 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644170 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644197 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644248 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644263 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644282 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644300 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644317 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644333 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644350 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644367 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644385 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644400 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644431 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644448 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644463 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644480 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644495 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644509 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644525 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644541 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644557 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644572 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644589 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644604 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644697 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644718 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644739 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644748 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644765 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644897 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644952 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.644995 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645035 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645079 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645101 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645120 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645123 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645173 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645216 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645256 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645297 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645370 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645383 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645427 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645508 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645549 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645592 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645664 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645803 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645848 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645887 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645931 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645969 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646010 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646054 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646095 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646134 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646174 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646215 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646255 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646329 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646370 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646455 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646498 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646541 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646580 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646653 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646692 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646733 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646773 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646815 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646919 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646966 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647008 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647089 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647131 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647171 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647211 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647250 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647292 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647333 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647374 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647458 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647503 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647545 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647585 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647652 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647696 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647737 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647779 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647819 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647862 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647904 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647943 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647981 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648022 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648064 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648107 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648149 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648212 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648253 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648293 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648342 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648395 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648439 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648479 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648520 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648562 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648612 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648680 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648725 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648770 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648813 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648854 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648911 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648992 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649042 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649084 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649129 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649175 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649260 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649310 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649357 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649402 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649451 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649542 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649587 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649684 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649733 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649782 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649823 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649867 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649910 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650034 5018 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650062 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650087 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650129 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650160 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650186 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650227 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650251 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650277 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650326 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650349 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650372 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650395 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650427 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650456 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650480 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650504 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650528 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650551 5018 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650573 5018 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650596 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650646 5018 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650670 5018 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650692 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650723 5018 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650747 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650771 5018 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650795 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650819 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650843 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650868 5018 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650892 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650923 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650947 5018 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650970 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650993 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651019 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651042 5018 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651065 5018 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651090 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651114 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651137 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651163 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651187 5018 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651213 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651237 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651261 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651283 5018 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651316 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651339 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651362 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651386 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651410 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651440 5018 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651467 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651491 5018 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651514 5018 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651537 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651563 5018 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651586 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651610 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651658 5018 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651686 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651714 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651743 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651770 5018 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651798 5018 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.654703 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.655718 5018 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645365 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645537 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.645662 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646200 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646223 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646250 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646341 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646521 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646893 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.646992 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.647479 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.658205 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648001 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648040 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648669 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.648607 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649083 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649115 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.649889 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650130 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.658263 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650149 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650305 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650771 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651106 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651166 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651197 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651403 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.651844 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.652108 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.652406 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.652851 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.654013 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.654386 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.654482 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.654692 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.655133 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.656336 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.656547 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.656564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.656937 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.657173 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.650376 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.658512 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.658977 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:04.158955557 +0000 UTC m=+20.743002274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.659397 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:04.159357838 +0000 UTC m=+20.743404545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.659827 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660018 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660231 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660426 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660527 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.658550 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.660583 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:04.160568782 +0000 UTC m=+20.744615539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660837 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660845 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.660851 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.661062 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.662862 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.663199 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.663778 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.663928 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.664002 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.664259 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.664284 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.664858 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.665218 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.665247 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.665309 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.665693 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.665924 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.666146 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.658537 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.666467 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.666762 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.666818 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.667219 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.668485 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.671761 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.672182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.672371 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.672396 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.672412 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.672416 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.672468 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:04.172450228 +0000 UTC m=+20.756496855 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.672765 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.672940 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.673138 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.674006 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.674291 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.674731 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.674326 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.674818 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.674838 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:03 crc kubenswrapper[5018]: E1014 06:50:03.674920 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:04.174890827 +0000 UTC m=+20.758937684 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.674717 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.675436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.675438 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.675640 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.675743 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677097 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677174 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677193 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677224 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677251 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677308 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677353 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.677461 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.681685 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.685075 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.685421 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.685842 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.685858 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.686029 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.686273 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.686605 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.686929 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.687033 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.688248 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.689949 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.690142 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.690918 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.690993 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.691077 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.691149 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.691440 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.691675 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.691881 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.692226 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.692246 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.693284 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.695591 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.696531 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.696766 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.697210 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.699514 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.705747 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.705977 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.706068 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.708535 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.708892 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.709135 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.719671 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.735979 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.741869 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.745422 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.745996 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.748249 5018 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100" exitCode=255 Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.748320 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100"} Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.756966 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757007 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757057 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757070 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757080 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757089 5018 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757097 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757106 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757114 5018 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757122 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757131 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757138 5018 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757148 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757156 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757165 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757173 5018 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757185 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757193 5018 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757201 5018 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757594 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757608 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757634 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757643 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757651 5018 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757659 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757668 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757676 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757684 5018 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757693 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757702 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757711 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757729 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757737 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757745 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757754 5018 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757762 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757770 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757778 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757786 5018 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757794 5018 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.757802 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758065 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758081 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758089 5018 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758098 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758106 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758115 5018 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758122 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758132 5018 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758140 5018 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758148 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758157 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758165 5018 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758173 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758181 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758929 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759039 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.758190 5018 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759156 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759213 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759223 5018 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759480 5018 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759499 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759508 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759518 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759526 5018 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759537 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759546 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759554 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759562 5018 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759674 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759683 5018 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759693 5018 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759702 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759711 5018 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759686 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759721 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.759976 5018 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760002 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760016 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760029 5018 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760039 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760051 5018 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760061 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760073 5018 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760085 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760097 5018 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760110 5018 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760120 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760130 5018 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760141 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760152 5018 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760164 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760175 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760186 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760199 5018 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760210 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760221 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760234 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760247 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760260 5018 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760271 5018 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760282 5018 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760293 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760304 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760315 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760326 5018 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760337 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760348 5018 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760360 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760371 5018 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760382 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760392 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760404 5018 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760415 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760426 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760436 5018 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760446 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760457 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760467 5018 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760478 5018 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760488 5018 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760498 5018 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760507 5018 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760517 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760527 5018 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760536 5018 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760546 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760555 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760565 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760574 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.760585 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.762375 5018 scope.go:117] "RemoveContainer" containerID="964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.764039 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.776463 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.791967 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.804126 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.832932 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.846791 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.850701 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.861043 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.876789 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 06:50:03 crc kubenswrapper[5018]: I1014 06:50:03.930306 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:50:03 crc kubenswrapper[5018]: W1014 06:50:03.946088 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-9cefa0be1e16e632c67bbee1ad844805410e4e327c021f759fad4105a9820c0b WatchSource:0}: Error finding container 9cefa0be1e16e632c67bbee1ad844805410e4e327c021f759fad4105a9820c0b: Status 404 returned error can't find the container with id 9cefa0be1e16e632c67bbee1ad844805410e4e327c021f759fad4105a9820c0b Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.058674 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.072738 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.073015 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.076152 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.088381 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.105599 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.118710 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.134246 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.145885 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.155890 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.164399 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.164525 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.164548 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:05.164523575 +0000 UTC m=+21.748570212 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.164577 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.164644 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.164691 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.164699 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:05.164687159 +0000 UTC m=+21.748733786 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.164732 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:05.16472188 +0000 UTC m=+21.748768507 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.165744 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.174173 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.185876 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.195754 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.204686 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.214773 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.226050 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.236485 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.265526 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.265679 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265794 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265838 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265854 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265915 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265957 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265981 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.265937 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:05.26591365 +0000 UTC m=+21.849960287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.266063 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:05.266050864 +0000 UTC m=+21.850097501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.608729 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.609511 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.610239 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.610953 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.612129 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.612776 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.613390 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.614889 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.615552 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.616802 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.617755 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.618921 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.619489 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.620067 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.621007 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.622682 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.625013 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.625993 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.627340 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.628771 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.629459 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.630110 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.630912 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.632109 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.633290 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.633898 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.636037 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.636948 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.638266 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.639124 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.640300 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.641080 5018 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.641244 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.644106 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.645038 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.645640 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.647853 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.649255 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.649964 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.650595 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.651485 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.652534 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.653947 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.654798 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.656359 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.657803 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.658579 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.659497 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.660771 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.661987 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.663194 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.663942 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.665311 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.666111 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.666810 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.666905 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.667646 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.691664 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.714653 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.736237 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.753465 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.753526 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4671e160f276308f28f24fe7bec64593ce77b411cc455858d5c5d5cdb02ca86c"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.755543 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.757973 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.758312 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.759444 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9cefa0be1e16e632c67bbee1ad844805410e4e327c021f759fad4105a9820c0b"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.759873 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.761831 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.761903 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.761921 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a67a0617f1d67b58b9c44368002e8f0b9465c5b104e7e9ff0cfa1e6c89c80d2b"} Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.775548 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.791093 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.808676 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.829156 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.843938 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.865316 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-db8xr"] Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.866010 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:04 crc kubenswrapper[5018]: W1014 06:50:04.867629 5018 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Oct 14 06:50:04 crc kubenswrapper[5018]: E1014 06:50:04.867677 5018 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.868434 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.890207 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.896683 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.930897 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.968570 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.972479 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/86d6c48b-2053-47ab-b68c-28442219d457-hosts-file\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.972702 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksfw4\" (UniqueName: \"kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.976656 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pwgtv"] Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.977158 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.983603 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.983657 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.983712 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 14 06:50:04 crc kubenswrapper[5018]: I1014 06:50:04.984109 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.015578 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.066839 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.073894 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlwr8\" (UniqueName: \"kubernetes.io/projected/696c1ace-2829-43a4-9103-a91c85ed06ad-kube-api-access-xlwr8\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.073952 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksfw4\" (UniqueName: \"kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.073976 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/86d6c48b-2053-47ab-b68c-28442219d457-hosts-file\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.074018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/696c1ace-2829-43a4-9103-a91c85ed06ad-serviceca\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.074037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/696c1ace-2829-43a4-9103-a91c85ed06ad-host\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.074340 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/86d6c48b-2053-47ab-b68c-28442219d457-hosts-file\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.085800 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.104273 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.117725 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.129839 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.146704 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.158690 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.170713 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175009 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175073 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/696c1ace-2829-43a4-9103-a91c85ed06ad-host\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175099 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlwr8\" (UniqueName: \"kubernetes.io/projected/696c1ace-2829-43a4-9103-a91c85ed06ad-kube-api-access-xlwr8\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175138 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175161 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175184 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/696c1ace-2829-43a4-9103-a91c85ed06ad-serviceca\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.175975 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/696c1ace-2829-43a4-9103-a91c85ed06ad-host\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.176034 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.176046 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:07.17601201 +0000 UTC m=+23.760058657 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.176142 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:07.176122283 +0000 UTC m=+23.760168910 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.176101 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.176215 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:07.176203015 +0000 UTC m=+23.760249652 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.177117 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/696c1ace-2829-43a4-9103-a91c85ed06ad-serviceca\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.197611 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlwr8\" (UniqueName: \"kubernetes.io/projected/696c1ace-2829-43a4-9103-a91c85ed06ad-kube-api-access-xlwr8\") pod \"node-ca-pwgtv\" (UID: \"696c1ace-2829-43a4-9103-a91c85ed06ad\") " pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.199126 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.213306 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.276541 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.276593 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276745 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276761 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276772 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276793 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276826 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:07.276808299 +0000 UTC m=+23.860854926 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276828 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276845 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.276900 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:07.276882651 +0000 UTC m=+23.860929328 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.287409 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pwgtv" Oct 14 06:50:05 crc kubenswrapper[5018]: W1014 06:50:05.298403 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod696c1ace_2829_43a4_9103_a91c85ed06ad.slice/crio-90e691aeb4d204e19d4929242e0bc2c1dc1ab58cd72b34a4e4d6569896fda2e1 WatchSource:0}: Error finding container 90e691aeb4d204e19d4929242e0bc2c1dc1ab58cd72b34a4e4d6569896fda2e1: Status 404 returned error can't find the container with id 90e691aeb4d204e19d4929242e0bc2c1dc1ab58cd72b34a4e4d6569896fda2e1 Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.604255 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.604301 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.604328 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.604374 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.604445 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:05 crc kubenswrapper[5018]: E1014 06:50:05.604524 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.764932 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pwgtv" event={"ID":"696c1ace-2829-43a4-9103-a91c85ed06ad","Type":"ContainerStarted","Data":"2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c"} Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.764979 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pwgtv" event={"ID":"696c1ace-2829-43a4-9103-a91c85ed06ad","Type":"ContainerStarted","Data":"90e691aeb4d204e19d4929242e0bc2c1dc1ab58cd72b34a4e4d6569896fda2e1"} Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.777124 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.788662 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.800463 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.808159 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-g7t5z"] Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.808678 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c2zpf"] Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.808842 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.811294 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fzdc9"] Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.811724 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.813312 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.813357 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.813777 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.814948 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-r7ftx"] Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.815161 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.820160 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.820212 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.820438 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.826167 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.826187 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.826232 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.826285 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827039 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827071 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827201 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827221 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827322 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827368 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827368 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827433 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827474 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.827545 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.836107 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.852317 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.871854 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.881968 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882014 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-proxy-tls\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882065 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8n2\" (UniqueName: \"kubernetes.io/projected/89e07068-0c9a-4200-b507-e98dca87d535-kube-api-access-cj8n2\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882088 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882113 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882140 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-multus-certs\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882198 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scgt6\" (UniqueName: \"kubernetes.io/projected/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-kube-api-access-scgt6\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882266 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882313 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-netns\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882394 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882444 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-bin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882473 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882494 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882516 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882546 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-system-cni-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882566 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cnibin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882594 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-rootfs\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882637 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882668 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882691 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882710 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cni-binary-copy\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882810 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882833 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882852 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-kubelet\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882913 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-socket-dir-parent\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882936 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-k8s-cni-cncf-io\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882966 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.882985 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883008 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5tp\" (UniqueName: \"kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883058 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883098 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-os-release\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883117 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-multus\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883138 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-daemon-config\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883180 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhkkx\" (UniqueName: \"kubernetes.io/projected/f1aa622b-96ac-42aa-8bd1-8fa174e65602-kube-api-access-nhkkx\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883223 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883330 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-cnibin\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883354 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883375 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-system-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883481 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-hostroot\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883520 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883546 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-binary-copy\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883567 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-conf-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883589 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-os-release\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.883607 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-etc-kubernetes\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.886684 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.899650 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.910378 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.923856 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.937288 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.952825 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.968662 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.984939 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.984975 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5tp\" (UniqueName: \"kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985011 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985031 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985046 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-daemon-config\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985061 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhkkx\" (UniqueName: \"kubernetes.io/projected/f1aa622b-96ac-42aa-8bd1-8fa174e65602-kube-api-access-nhkkx\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985097 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985120 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985135 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-os-release\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985150 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-multus\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985190 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985176 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985222 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985272 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-cnibin\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985212 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-cnibin\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985270 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985330 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-multus\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985366 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-system-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985397 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-hostroot\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985409 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985436 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985470 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-binary-copy\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985499 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-conf-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985525 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-os-release\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985529 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-os-release\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985581 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-etc-kubernetes\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985607 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985658 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-proxy-tls\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985683 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985709 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8n2\" (UniqueName: \"kubernetes.io/projected/89e07068-0c9a-4200-b507-e98dca87d535-kube-api-access-cj8n2\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985744 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-system-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985746 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985802 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985833 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-multus-certs\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985865 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-netns\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985903 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scgt6\" (UniqueName: \"kubernetes.io/projected/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-kube-api-access-scgt6\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985909 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-etc-kubernetes\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985935 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985992 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986023 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-bin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986046 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-system-cni-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986057 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986070 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cnibin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986120 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cnibin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986145 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986155 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986169 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986191 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986197 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-daemon-config\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986225 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986227 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-cni-bin\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986248 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-system-cni-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986195 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986260 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986283 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-rootfs\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986307 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-rootfs\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986318 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986331 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986342 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986346 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985868 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-cni-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.985608 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-os-release\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986459 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986465 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-netns\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986483 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986500 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cni-binary-copy\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986518 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-kubelet\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986533 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986576 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986592 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-k8s-cni-cncf-io\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986637 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986648 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986654 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-socket-dir-parent\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986682 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-socket-dir-parent\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986726 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986747 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-hostroot\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986873 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986916 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-k8s-cni-cncf-io\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986942 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.986993 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987098 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e07068-0c9a-4200-b507-e98dca87d535-cni-binary-copy\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987114 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-run-multus-certs\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987142 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-multus-conf-dir\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987171 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1aa622b-96ac-42aa-8bd1-8fa174e65602-host-var-lib-kubelet\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987200 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987529 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1aa622b-96ac-42aa-8bd1-8fa174e65602-cni-binary-copy\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.987674 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-mcd-auth-proxy-config\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.988496 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e07068-0c9a-4200-b507-e98dca87d535-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.989055 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.989850 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:05 crc kubenswrapper[5018]: I1014 06:50:05.989948 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-proxy-tls\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.003374 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhkkx\" (UniqueName: \"kubernetes.io/projected/f1aa622b-96ac-42aa-8bd1-8fa174e65602-kube-api-access-nhkkx\") pod \"multus-r7ftx\" (UID: \"f1aa622b-96ac-42aa-8bd1-8fa174e65602\") " pod="openshift-multus/multus-r7ftx" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.009222 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scgt6\" (UniqueName: \"kubernetes.io/projected/124fa7dd-4f8d-4d66-9223-f6bcc6df7242-kube-api-access-scgt6\") pod \"machine-config-daemon-g7t5z\" (UID: \"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\") " pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.009819 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.012313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8n2\" (UniqueName: \"kubernetes.io/projected/89e07068-0c9a-4200-b507-e98dca87d535-kube-api-access-cj8n2\") pod \"multus-additional-cni-plugins-fzdc9\" (UID: \"89e07068-0c9a-4200-b507-e98dca87d535\") " pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.012926 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5tp\" (UniqueName: \"kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp\") pod \"ovnkube-node-c2zpf\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.023984 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.037235 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.046959 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.061756 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.079244 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: E1014 06:50:06.090001 5018 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 14 06:50:06 crc kubenswrapper[5018]: E1014 06:50:06.090072 5018 projected.go:194] Error preparing data for projected volume kube-api-access-ksfw4 for pod openshift-dns/node-resolver-db8xr: failed to sync configmap cache: timed out waiting for the condition Oct 14 06:50:06 crc kubenswrapper[5018]: E1014 06:50:06.090146 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4 podName:86d6c48b-2053-47ab-b68c-28442219d457 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:06.590122554 +0000 UTC m=+23.174169171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ksfw4" (UniqueName: "kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4") pod "node-resolver-db8xr" (UID: "86d6c48b-2053-47ab-b68c-28442219d457") : failed to sync configmap cache: timed out waiting for the condition Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.091169 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.113410 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.126059 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.127227 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.134432 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.141268 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-r7ftx" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.145954 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.147752 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" Oct 14 06:50:06 crc kubenswrapper[5018]: W1014 06:50:06.168288 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89e07068_0c9a_4200_b507_e98dca87d535.slice/crio-08647e8151042e9a08a89fe26ceb43d4a060437ccf356ac7b1272afdc2e74071 WatchSource:0}: Error finding container 08647e8151042e9a08a89fe26ceb43d4a060437ccf356ac7b1272afdc2e74071: Status 404 returned error can't find the container with id 08647e8151042e9a08a89fe26ceb43d4a060437ccf356ac7b1272afdc2e74071 Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.390381 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.593641 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksfw4\" (UniqueName: \"kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.599949 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksfw4\" (UniqueName: \"kubernetes.io/projected/86d6c48b-2053-47ab-b68c-28442219d457-kube-api-access-ksfw4\") pod \"node-resolver-db8xr\" (UID: \"86d6c48b-2053-47ab-b68c-28442219d457\") " pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.679306 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-db8xr" Oct 14 06:50:06 crc kubenswrapper[5018]: W1014 06:50:06.741495 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86d6c48b_2053_47ab_b68c_28442219d457.slice/crio-057cbe39bab17fd5a235492bd187c8ce7ecb5b1002647e60aaa5a19d84f56657 WatchSource:0}: Error finding container 057cbe39bab17fd5a235492bd187c8ce7ecb5b1002647e60aaa5a19d84f56657: Status 404 returned error can't find the container with id 057cbe39bab17fd5a235492bd187c8ce7ecb5b1002647e60aaa5a19d84f56657 Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.772047 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.775707 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerStarted","Data":"e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.775784 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerStarted","Data":"046e661dd48e6cc60c6a30ffddfba5213e6d8b072105482d1aee5bd04f01d0ad"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.777011 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-db8xr" event={"ID":"86d6c48b-2053-47ab-b68c-28442219d457","Type":"ContainerStarted","Data":"057cbe39bab17fd5a235492bd187c8ce7ecb5b1002647e60aaa5a19d84f56657"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.781103 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31" exitCode=0 Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.781209 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.781301 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerStarted","Data":"08647e8151042e9a08a89fe26ceb43d4a060437ccf356ac7b1272afdc2e74071"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.791485 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.791554 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.791571 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"8793564373cc14fe74fe85526d8c785bf3fa4746000f08dfb68cd49cf4a158e8"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.793971 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" exitCode=0 Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.794010 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.794062 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"f8b8d9748c40989728b53957cb281152cf9134431fc51993c123e5594ee19121"} Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.798281 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.816281 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.833012 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.851570 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.867175 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.879407 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.921896 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.941896 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.978502 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:06 crc kubenswrapper[5018]: I1014 06:50:06.991291 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:06Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.017005 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.031324 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.042142 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.057879 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.075358 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.087265 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.106020 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.136378 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.150011 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.167477 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.187150 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.200839 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.201087 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.201232 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:11.201210955 +0000 UTC m=+27.785257582 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.201306 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.201331 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.201467 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.201551 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:11.201532575 +0000 UTC m=+27.785579202 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.201554 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.201601 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:11.201595386 +0000 UTC m=+27.785642013 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.222895 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.239152 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.255613 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.275360 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.293878 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.302276 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.302333 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302533 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302561 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302555 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302610 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302577 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302656 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302728 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:11.302700744 +0000 UTC m=+27.886747371 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.302755 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:11.302745445 +0000 UTC m=+27.886792072 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.311431 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.604312 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.604394 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.604462 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.604508 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.604719 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:07 crc kubenswrapper[5018]: E1014 06:50:07.604847 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.799315 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.799889 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.799910 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.801118 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-db8xr" event={"ID":"86d6c48b-2053-47ab-b68c-28442219d457","Type":"ContainerStarted","Data":"b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa"} Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.804114 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524" exitCode=0 Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.804185 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524"} Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.823418 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.843832 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.862695 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.890203 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.909492 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.925194 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.938676 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.952504 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.966313 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.983605 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:07 crc kubenswrapper[5018]: I1014 06:50:07.996511 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:07Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.019781 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.037468 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.056350 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.073198 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.097347 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.114428 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.128871 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.142279 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.159091 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.174806 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.192091 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.205933 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.223305 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.238836 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.261114 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.278946 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.292951 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.814563 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.814642 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.814658 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.818889 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68" exitCode=0 Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.818951 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68"} Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.841652 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.857883 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.873949 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.895558 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.909975 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.924980 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.938993 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.963438 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.981179 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:08 crc kubenswrapper[5018]: I1014 06:50:08.996874 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:08Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.014639 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.029257 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.046113 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.059963 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.604316 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.604377 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.604418 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.604510 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.604611 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.604808 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.818486 5018 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.820996 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.821039 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.821058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.821207 5018 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.827013 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77" exitCode=0 Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.827081 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.835775 5018 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.836055 5018 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.837807 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.837882 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.837912 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.837948 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.837972 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.863955 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.864302 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.869448 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.869510 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.869530 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.869556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.869572 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.884125 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.891185 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.895661 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.895723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.895743 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.895771 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.895790 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.906218 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.914568 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.918540 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.918604 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.918647 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.918676 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.918696 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.922979 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.933964 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.937951 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.938013 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.938030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.938056 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.938070 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.945602 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.953186 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: E1014 06:50:09.953749 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.961580 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.961663 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.961681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.961706 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.961725 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:09Z","lastTransitionTime":"2025-10-14T06:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.967654 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.981927 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:09 crc kubenswrapper[5018]: I1014 06:50:09.997433 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:09Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.010991 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.030475 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.055072 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.064863 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.064908 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.064935 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.064957 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.064969 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.075137 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.094760 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.108722 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.169183 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.169223 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.169236 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.169253 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.169266 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.273464 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.273970 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.273989 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.274015 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.274030 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.377536 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.377580 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.377589 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.377607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.377634 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.480778 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.480858 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.480885 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.480915 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.480940 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.584153 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.584214 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.584232 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.584259 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.584281 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.687454 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.687542 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.687563 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.687596 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.687655 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.791268 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.791343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.791361 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.791387 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.791404 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.834657 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342" exitCode=0 Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.834779 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.842805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.861002 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.877867 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.894897 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.897484 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.897565 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.897588 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.897616 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.897691 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:10Z","lastTransitionTime":"2025-10-14T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.919224 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.942924 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.960325 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:10 crc kubenswrapper[5018]: I1014 06:50:10.981161 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:10Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.000017 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.000093 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.000120 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.000160 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.000189 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.012460 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.028785 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.044722 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.058599 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.078890 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.102335 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.103060 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.103126 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.103145 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.103173 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.103193 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.124126 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.206051 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.206109 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.206121 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.206143 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.206157 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.256160 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.256344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.256407 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:19.25637381 +0000 UTC m=+35.840420477 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.256446 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.256451 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.256510 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:19.256490363 +0000 UTC m=+35.840537030 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.256590 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.256692 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:19.256676728 +0000 UTC m=+35.840723395 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.310273 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.310352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.310373 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.310396 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.310409 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.358037 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.358161 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358309 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358355 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358376 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358492 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:19.358464205 +0000 UTC m=+35.942510872 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358526 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358570 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358599 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.358771 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:19.358733593 +0000 UTC m=+35.942780410 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.414054 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.414109 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.414121 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.414147 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.414160 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.516709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.516775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.516792 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.516819 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.516839 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.604775 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.604844 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.604844 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.605019 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.605151 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:11 crc kubenswrapper[5018]: E1014 06:50:11.605350 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.619998 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.620073 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.620098 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.620135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.620162 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.724988 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.725068 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.725085 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.725115 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.725140 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.827696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.827764 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.827783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.827811 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.827833 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.856277 5018 generic.go:334] "Generic (PLEG): container finished" podID="89e07068-0c9a-4200-b507-e98dca87d535" containerID="ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca" exitCode=0 Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.857315 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerDied","Data":"ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.879306 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.895954 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.918170 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.930839 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.930902 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.930924 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.930956 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.930979 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:11Z","lastTransitionTime":"2025-10-14T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.940038 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.957896 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.973670 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:11 crc kubenswrapper[5018]: I1014 06:50:11.986223 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:11Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.005995 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.021662 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.034665 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.034704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.034718 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.034739 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.034754 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.040653 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.070561 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.091141 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.108981 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.125102 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.137754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.137811 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.137827 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.137852 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.137871 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.240858 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.240949 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.240979 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.241026 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.241054 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.344693 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.344741 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.344758 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.344781 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.344801 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.456770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.457384 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.457416 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.457475 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.457501 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.561253 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.561334 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.561348 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.561374 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.561388 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.663658 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.663724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.663744 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.663770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.663788 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.767284 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.767332 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.767346 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.767366 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.767382 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.867984 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" event={"ID":"89e07068-0c9a-4200-b507-e98dca87d535","Type":"ContainerStarted","Data":"8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.870467 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.870518 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.870536 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.870610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.870666 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.879257 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.879918 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.892852 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.913027 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.918307 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.933241 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.949962 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.967003 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.973268 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.973306 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.973319 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.973339 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.973353 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:12Z","lastTransitionTime":"2025-10-14T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:12 crc kubenswrapper[5018]: I1014 06:50:12.988882 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:12Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.009670 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.038686 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.063740 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.076197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.076253 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.076268 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.076292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.076310 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.085429 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.104155 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.120934 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.142584 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.157497 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.177492 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.179258 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.179320 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.179335 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.179360 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.179375 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.196428 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.214142 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.233195 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.248867 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.269870 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.282459 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.282534 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.282547 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.282570 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.282584 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.296342 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.318121 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.346434 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.369333 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.385122 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.385311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.385423 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.385517 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.385595 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.390164 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.417243 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.460897 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.488519 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.489640 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.489672 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.489685 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.489702 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.489712 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.592449 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.592550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.592571 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.592595 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.592612 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.604859 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.604863 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:13 crc kubenswrapper[5018]: E1014 06:50:13.605054 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:13 crc kubenswrapper[5018]: E1014 06:50:13.605199 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.604872 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:13 crc kubenswrapper[5018]: E1014 06:50:13.605417 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.696016 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.696126 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.696151 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.696180 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.696199 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.799608 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.799687 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.799698 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.799721 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.799734 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.883286 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.884079 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.903209 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.903294 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.903309 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.903337 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.903355 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:13Z","lastTransitionTime":"2025-10-14T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.969233 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:13 crc kubenswrapper[5018]: I1014 06:50:13.988738 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.002454 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:13Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.007247 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.007307 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.007326 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.007353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.007372 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.016924 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.031794 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.047492 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.061249 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.076094 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.091921 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.105540 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.110325 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.110392 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.110408 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.110431 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.110447 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.121022 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.135462 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.156537 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.173007 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.194064 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.213768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.213818 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.213831 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.213855 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.213870 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.317023 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.317086 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.317102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.317124 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.317139 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.420918 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.420993 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.421007 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.421036 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.421057 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.525681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.525750 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.525771 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.525801 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.525820 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.629769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.629934 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.629964 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.630045 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.630113 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.636877 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.650582 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.669703 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.692411 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.715282 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.732753 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.732808 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.732748 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.732821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.733002 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.733021 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.750439 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.765176 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.783470 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.804495 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.823808 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.836358 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.836414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.836428 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.836453 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.836468 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.853063 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.869248 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.885075 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:14Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.886537 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.939776 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.939868 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.939884 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.939913 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:14 crc kubenswrapper[5018]: I1014 06:50:14.939935 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:14Z","lastTransitionTime":"2025-10-14T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.069262 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.069312 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.069321 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.069337 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.069348 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.172024 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.172067 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.172079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.172095 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.172104 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.275457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.275506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.275517 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.275540 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.275555 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.378379 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.378415 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.378427 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.378444 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.378456 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.480471 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.480531 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.480550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.480574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.480591 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.583895 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.583934 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.583943 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.583957 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.583966 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.604612 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.604730 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.604919 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:15 crc kubenswrapper[5018]: E1014 06:50:15.604914 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:15 crc kubenswrapper[5018]: E1014 06:50:15.605004 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:15 crc kubenswrapper[5018]: E1014 06:50:15.605067 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.686758 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.686828 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.686841 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.686859 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.686871 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.790595 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.790712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.790739 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.790772 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.790798 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.892764 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.893705 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.893726 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.893770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.893782 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.893027 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/0.log" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.898124 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2" exitCode=1 Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.898180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2"} Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.898947 5018 scope.go:117] "RemoveContainer" containerID="a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.919781 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:15Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.943707 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:15Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.962191 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:15Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.984749 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:15Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:15.580602 6333 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:15.580684 6333 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:15.580765 6333 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:15.580789 6333 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:15.580831 6333 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:15.580857 6333 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:15.580880 6333 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:15.580897 6333 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:15.580936 6333 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:15.580936 6333 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:15.580959 6333 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:15.580970 6333 factory.go:656] Stopping watch factory\\\\nI1014 06:50:15.580981 6333 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:15.580985 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:15.580992 6333 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:15Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.997288 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.997316 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.997324 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.997336 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:15 crc kubenswrapper[5018]: I1014 06:50:15.997346 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:15Z","lastTransitionTime":"2025-10-14T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.007032 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.022610 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.039103 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.056385 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.069926 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.087681 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.100452 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.100499 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.100516 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.100541 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.100559 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.104608 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.115457 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.128059 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.152996 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.205170 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.205403 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.205414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.205431 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.205443 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.307723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.307751 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.307761 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.307775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.307786 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.411134 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.411364 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.411552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.411697 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.411788 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.514167 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.514412 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.514567 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.514732 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.514885 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.617234 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.617492 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.617685 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.617791 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.617885 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.719826 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.720045 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.720152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.720242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.720323 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.827116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.827162 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.827179 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.827199 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.827249 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.908744 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/0.log" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.913524 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.913748 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.931304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.931372 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.931396 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.931426 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.931452 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:16Z","lastTransitionTime":"2025-10-14T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.944008 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.959691 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.978673 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:16 crc kubenswrapper[5018]: I1014 06:50:16.995107 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:16Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.018106 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.034093 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.034132 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.034140 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.034154 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.034163 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.041226 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.055377 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.068218 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.084298 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.107833 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.129339 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.136412 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.136478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.136502 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.136532 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.136557 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.163855 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:15Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:15.580602 6333 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:15.580684 6333 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:15.580765 6333 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:15.580789 6333 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:15.580831 6333 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:15.580857 6333 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:15.580880 6333 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:15.580897 6333 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:15.580936 6333 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:15.580936 6333 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:15.580959 6333 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:15.580970 6333 factory.go:656] Stopping watch factory\\\\nI1014 06:50:15.580981 6333 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:15.580985 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:15.580992 6333 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.183686 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.202319 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.239659 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.239707 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.239720 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.239737 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.239752 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.342714 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.342778 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.342801 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.342833 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.342856 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.445479 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.445550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.445568 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.445596 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.445620 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.548933 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.548985 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.549001 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.549027 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.549079 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.604473 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.604593 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:17 crc kubenswrapper[5018]: E1014 06:50:17.604823 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.604912 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:17 crc kubenswrapper[5018]: E1014 06:50:17.604994 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:17 crc kubenswrapper[5018]: E1014 06:50:17.605242 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.652686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.652739 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.652756 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.652780 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.652798 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.755690 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.755737 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.755753 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.755776 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.755793 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.858566 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.858698 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.858723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.858752 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.858771 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.920439 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/1.log" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.921400 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/0.log" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.926341 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" exitCode=1 Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.926398 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.926511 5018 scope.go:117] "RemoveContainer" containerID="a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.927589 5018 scope.go:117] "RemoveContainer" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" Oct 14 06:50:17 crc kubenswrapper[5018]: E1014 06:50:17.927927 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.953837 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.961527 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.961582 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.961601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.961634 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.961714 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:17Z","lastTransitionTime":"2025-10-14T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.977437 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:17 crc kubenswrapper[5018]: I1014 06:50:17.997171 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:17Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.032940 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:15Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:15.580602 6333 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:15.580684 6333 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:15.580765 6333 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:15.580789 6333 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:15.580831 6333 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:15.580857 6333 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:15.580880 6333 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:15.580897 6333 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:15.580936 6333 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:15.580936 6333 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:15.580959 6333 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:15.580970 6333 factory.go:656] Stopping watch factory\\\\nI1014 06:50:15.580981 6333 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:15.580985 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:15.580992 6333 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.051915 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.065413 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.065480 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.065502 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.065531 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.065553 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.071525 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.086365 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.104330 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.118571 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.135044 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.151863 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.166076 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.167457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.167494 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.167503 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.167517 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.167526 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.180070 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.189148 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.246229 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.261356 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.270359 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.270393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.270405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.270421 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.270432 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.274290 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.292784 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.308465 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.328357 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.346744 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.363453 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.372781 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.372821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.372832 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.372848 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.372860 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.393375 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:15Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:15.580602 6333 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:15.580684 6333 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:15.580765 6333 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:15.580789 6333 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:15.580831 6333 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:15.580857 6333 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:15.580880 6333 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:15.580897 6333 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:15.580936 6333 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:15.580936 6333 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:15.580959 6333 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:15.580970 6333 factory.go:656] Stopping watch factory\\\\nI1014 06:50:15.580981 6333 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:15.580985 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:15.580992 6333 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.409764 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.423979 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.438428 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.455264 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.473181 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.474747 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.474793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.474806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.474826 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.474839 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.482970 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt"] Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.483596 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.485955 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.486981 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.493485 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.508732 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.521143 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.535322 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/898b50bd-ba6a-481e-8337-11aaa1d75199-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.535484 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.535553 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcdjk\" (UniqueName: \"kubernetes.io/projected/898b50bd-ba6a-481e-8337-11aaa1d75199-kube-api-access-dcdjk\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.535756 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.536142 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.558788 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.576461 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.577519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.577576 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.577595 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.577657 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.577684 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.600988 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.624647 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.637196 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.637302 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/898b50bd-ba6a-481e-8337-11aaa1d75199-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.637409 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.637465 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcdjk\" (UniqueName: \"kubernetes.io/projected/898b50bd-ba6a-481e-8337-11aaa1d75199-kube-api-access-dcdjk\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.638343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.639035 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/898b50bd-ba6a-481e-8337-11aaa1d75199-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.646534 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/898b50bd-ba6a-481e-8337-11aaa1d75199-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.659180 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c917eea1019020495244a95a70d75311708e0ae0a04a20dec5a5fb7a002ee2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:15Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:15.580602 6333 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:15.580684 6333 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:15.580765 6333 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:15.580789 6333 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:15.580831 6333 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:15.580857 6333 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:15.580880 6333 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:15.580897 6333 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:15.580936 6333 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:15.580936 6333 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:15.580959 6333 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:15.580970 6333 factory.go:656] Stopping watch factory\\\\nI1014 06:50:15.580981 6333 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:15.580985 6333 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:15.580992 6333 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.666838 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcdjk\" (UniqueName: \"kubernetes.io/projected/898b50bd-ba6a-481e-8337-11aaa1d75199-kube-api-access-dcdjk\") pod \"ovnkube-control-plane-749d76644c-mxxzt\" (UID: \"898b50bd-ba6a-481e-8337-11aaa1d75199\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.677447 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.680196 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.680255 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.680274 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.680297 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.680314 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.697710 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.715193 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.729280 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.747245 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.768749 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.783563 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.783611 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.783657 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.783681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.783701 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.789707 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.797722 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.887175 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.887226 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.887242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.887261 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.887276 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.931379 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" event={"ID":"898b50bd-ba6a-481e-8337-11aaa1d75199","Type":"ContainerStarted","Data":"fa350c8dcb0c33aeedfd407c359a183a3d8dd629b8be0f1b95cb0109136b05f1"} Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.934543 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/1.log" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.938661 5018 scope.go:117] "RemoveContainer" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" Oct 14 06:50:18 crc kubenswrapper[5018]: E1014 06:50:18.938817 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.963034 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.986468 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:18Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.990789 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.990898 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.990925 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.990955 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:18 crc kubenswrapper[5018]: I1014 06:50:18.990979 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:18Z","lastTransitionTime":"2025-10-14T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.005703 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.022746 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.040757 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.059150 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.081512 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.092264 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.093797 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.093851 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.093869 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.093892 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.093903 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.105010 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.122130 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.136504 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.151215 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.181532 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.197043 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.197110 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.197135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.197166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.197187 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.201231 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.221020 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.300265 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.300303 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.300312 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.300328 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.300338 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.347252 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.347371 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.347353371 +0000 UTC m=+51.931399998 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.347570 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.347596 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.347694 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.347738 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.347730622 +0000 UTC m=+51.931777239 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.347914 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.348062 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.3480154 +0000 UTC m=+51.932062077 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.402829 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.402867 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.402881 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.402896 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.402906 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.449720 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.449790 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.449978 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450004 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450023 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450091 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.450066534 +0000 UTC m=+52.034113181 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450559 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450584 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450597 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.450664 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.45063087 +0000 UTC m=+52.034677517 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.505826 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.505870 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.505884 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.505902 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.505916 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.604967 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.605097 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.605133 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.605257 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.605449 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.605545 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.608800 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.608854 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.608873 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.608897 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.608914 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.635947 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gl9fg"] Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.636613 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.636733 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.651981 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.652222 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xwjj\" (UniqueName: \"kubernetes.io/projected/2530b46d-5f3a-4495-af23-3ad85327c99c-kube-api-access-7xwjj\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.654297 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.678688 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.693537 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.709252 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.711513 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.711544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.711557 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.711575 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.711590 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.724741 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.736651 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.753742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xwjj\" (UniqueName: \"kubernetes.io/projected/2530b46d-5f3a-4495-af23-3ad85327c99c-kube-api-access-7xwjj\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.753866 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.754027 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: E1014 06:50:19.754112 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:20.254084916 +0000 UTC m=+36.838131583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.756770 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.769077 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.783805 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xwjj\" (UniqueName: \"kubernetes.io/projected/2530b46d-5f3a-4495-af23-3ad85327c99c-kube-api-access-7xwjj\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.789116 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.813310 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.814077 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.814120 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.814133 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.814153 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.814164 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.830300 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.848947 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.871079 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.884088 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.903326 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.916238 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.916317 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.916336 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.916361 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.916379 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:19Z","lastTransitionTime":"2025-10-14T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.926405 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.943180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" event={"ID":"898b50bd-ba6a-481e-8337-11aaa1d75199","Type":"ContainerStarted","Data":"9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.943244 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" event={"ID":"898b50bd-ba6a-481e-8337-11aaa1d75199","Type":"ContainerStarted","Data":"61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7"} Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.965431 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:19 crc kubenswrapper[5018]: I1014 06:50:19.997159 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:19Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.016890 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.018745 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.018800 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.018816 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.018837 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.018853 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.036384 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.056051 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.073270 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.090255 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.106178 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.121752 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.122201 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.122265 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.122290 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.122323 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.122348 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.144447 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.159972 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.177572 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.190611 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.206387 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.223216 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.224543 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.224590 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.224606 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.224633 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.224703 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.239551 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.258319 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.258484 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.258549 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:21.258532022 +0000 UTC m=+37.842578659 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.287296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.287367 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.287387 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.287414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.287432 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.303432 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.307967 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.308040 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.308061 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.308119 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.308142 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.323805 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.327849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.327871 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.327882 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.327898 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.327911 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.345063 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.349352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.349393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.349407 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.349424 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.349439 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.365020 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.369513 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.369585 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.369609 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.369681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.369708 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.384997 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:20Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:20 crc kubenswrapper[5018]: E1014 06:50:20.385235 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.387318 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.387353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.387365 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.387381 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.387393 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.490338 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.490398 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.490422 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.490450 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.490471 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.593411 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.593481 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.593509 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.593541 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.593607 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.696028 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.696060 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.696071 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.696086 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.696097 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.799171 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.799273 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.799550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.799587 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.799610 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.902543 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.902696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.902719 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.902746 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:20 crc kubenswrapper[5018]: I1014 06:50:20.902767 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:20Z","lastTransitionTime":"2025-10-14T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.006327 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.006398 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.006421 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.006451 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.006473 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.109160 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.109251 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.109269 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.109296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.109320 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.212463 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.212532 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.212550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.212574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.212593 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.269397 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.269574 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.269657 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:23.269619977 +0000 UTC m=+39.853666614 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.315806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.315878 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.315897 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.315969 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.315990 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.419410 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.419463 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.419476 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.419497 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.419510 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.523208 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.523292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.523311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.523338 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.523361 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.604811 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.604840 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.604841 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.605009 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.605183 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.605334 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.605464 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:21 crc kubenswrapper[5018]: E1014 06:50:21.605707 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.627191 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.627262 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.627284 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.627340 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.627363 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.730979 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.731051 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.731073 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.731104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.731125 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.834227 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.834324 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.834346 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.834374 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.834398 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.937791 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.937840 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.937853 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.937872 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:21 crc kubenswrapper[5018]: I1014 06:50:21.937885 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:21Z","lastTransitionTime":"2025-10-14T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.040899 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.040971 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.040995 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.041026 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.041050 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.145178 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.145262 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.145287 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.145318 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.145342 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.249158 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.249252 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.249271 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.249294 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.249312 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.352782 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.353101 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.353118 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.353148 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.353166 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.455480 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.455565 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.455587 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.455616 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.455691 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.558308 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.558361 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.558380 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.558398 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.558411 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.661769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.661818 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.661829 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.661849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.661862 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.765434 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.765506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.765527 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.765556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.765576 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.868562 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.868610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.868662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.868688 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.868707 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.970819 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.970878 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.970896 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.970921 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:22 crc kubenswrapper[5018]: I1014 06:50:22.970938 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:22Z","lastTransitionTime":"2025-10-14T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.075287 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.075363 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.075382 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.075406 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.075424 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.178783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.178855 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.178874 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.178897 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.178915 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.281372 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.281436 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.281447 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.281461 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.281472 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.293049 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.293198 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.293253 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:27.293238227 +0000 UTC m=+43.877284854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.384006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.384095 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.384116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.384141 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.384158 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.486802 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.486875 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.486887 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.486957 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.486969 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.593123 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.593191 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.593208 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.593232 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.593250 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.604660 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.604676 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.604724 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.604785 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.604995 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.605070 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.605150 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.605301 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.696380 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.696443 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.696461 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.696487 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.696504 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.760867 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.762661 5018 scope.go:117] "RemoveContainer" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" Oct 14 06:50:23 crc kubenswrapper[5018]: E1014 06:50:23.762971 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.800293 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.800346 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.800358 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.800393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.800407 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.902650 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.902721 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.902744 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.902773 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:23 crc kubenswrapper[5018]: I1014 06:50:23.902797 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:23Z","lastTransitionTime":"2025-10-14T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.006198 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.006304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.006321 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.006345 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.006366 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.109329 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.109428 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.109447 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.109472 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.109488 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.212565 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.212610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.212678 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.212709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.212728 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.315519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.315581 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.315594 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.315613 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.315647 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.418063 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.418117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.418134 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.418160 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.418178 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.520937 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.520991 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.521008 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.521030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.521047 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.624408 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.624470 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.624487 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.624509 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.624527 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.628070 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.649252 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.668356 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.683527 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.704073 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.719475 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.727216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.727287 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.727313 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.727343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.727367 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.742039 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.766385 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.786286 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.807796 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.830513 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.830592 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.830615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.830670 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.830691 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.837882 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.858487 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.888978 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.910988 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.934575 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.934687 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.934714 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.934748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.934771 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:24Z","lastTransitionTime":"2025-10-14T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.936798 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:24 crc kubenswrapper[5018]: I1014 06:50:24.954823 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:24Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.037420 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.037497 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.037517 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.037543 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.037561 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.141042 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.141102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.141124 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.141150 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.141167 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.245280 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.245357 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.245396 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.245433 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.245458 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.347793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.347850 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.347866 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.347889 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.347909 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.451165 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.451226 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.451243 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.451264 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.451278 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.554724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.554789 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.554806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.554832 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.554849 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.604748 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.604803 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.604858 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.604748 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:25 crc kubenswrapper[5018]: E1014 06:50:25.604934 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:25 crc kubenswrapper[5018]: E1014 06:50:25.605059 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:25 crc kubenswrapper[5018]: E1014 06:50:25.605285 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:25 crc kubenswrapper[5018]: E1014 06:50:25.605383 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.658352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.658460 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.658477 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.658501 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.658520 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.761741 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.761800 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.761824 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.761849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.761867 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.864729 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.864796 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.864819 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.864847 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.864868 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.967605 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.967676 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.967689 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.967748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:25 crc kubenswrapper[5018]: I1014 06:50:25.967762 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:25Z","lastTransitionTime":"2025-10-14T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.071613 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.071707 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.071724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.071747 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.071764 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.174571 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.174607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.174618 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.174662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.174677 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.276731 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.276781 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.276798 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.276822 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.276838 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.379389 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.379437 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.379455 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.379478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.379494 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.483180 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.483245 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.483267 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.483293 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.483312 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.586000 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.586074 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.586099 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.586129 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.586153 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.689345 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.689396 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.689414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.689439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.689457 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.793007 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.793079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.793102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.793129 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.793150 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.895938 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.895987 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.895999 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.896017 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.896031 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.998914 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.999026 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.999054 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.999084 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:26 crc kubenswrapper[5018]: I1014 06:50:26.999106 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:26Z","lastTransitionTime":"2025-10-14T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.102208 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.102279 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.102301 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.102331 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.102353 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.206386 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.206456 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.206473 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.206496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.206515 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.309336 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.309406 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.309422 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.309449 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.309466 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.340565 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.340866 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.340998 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:35.340962432 +0000 UTC m=+51.925009109 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.413016 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.413083 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.413104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.413128 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.413146 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.516485 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.516538 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.516556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.516579 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.516596 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.604776 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.604803 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.604810 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.604892 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.604956 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.605034 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.605073 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:27 crc kubenswrapper[5018]: E1014 06:50:27.605146 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.619094 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.619133 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.619144 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.619162 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.619179 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.722601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.722713 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.722737 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.722773 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.722798 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.826489 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.826570 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.826589 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.826615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.826698 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.929560 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.929645 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.929664 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.929688 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:27 crc kubenswrapper[5018]: I1014 06:50:27.929705 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:27Z","lastTransitionTime":"2025-10-14T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.032806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.032855 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.032872 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.032898 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.032917 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.135747 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.135811 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.135826 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.135845 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.135860 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.238726 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.238787 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.238802 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.238818 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.238828 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.346858 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.346903 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.346915 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.346933 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.346946 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.449951 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.450008 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.450023 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.450046 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.450061 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.552974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.553042 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.553067 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.553096 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.553118 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.656271 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.656332 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.656351 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.656375 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.656393 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.764924 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.764962 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.764986 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.765005 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.765018 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.868781 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.868849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.868868 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.868893 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.868912 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.971754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.971833 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.971858 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.971892 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:28 crc kubenswrapper[5018]: I1014 06:50:28.971920 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:28Z","lastTransitionTime":"2025-10-14T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.075300 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.075344 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.075353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.075368 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.075377 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.179237 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.179337 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.179367 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.179397 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.179424 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.282304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.282379 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.282398 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.282424 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.282443 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.384996 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.385096 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.385122 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.385152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.385174 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.488718 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.488781 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.488799 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.488824 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.488844 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.592044 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.592116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.592135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.592164 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.592187 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.604500 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.604570 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:29 crc kubenswrapper[5018]: E1014 06:50:29.604773 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.604843 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.604846 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:29 crc kubenswrapper[5018]: E1014 06:50:29.605008 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:29 crc kubenswrapper[5018]: E1014 06:50:29.605894 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:29 crc kubenswrapper[5018]: E1014 06:50:29.606010 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.694472 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.694532 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.694550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.694574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.694591 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.797763 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.797899 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.797925 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.797954 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.797978 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.901547 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.901598 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.901615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.901672 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:29 crc kubenswrapper[5018]: I1014 06:50:29.901689 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:29Z","lastTransitionTime":"2025-10-14T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.004102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.004154 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.004173 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.004200 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.004222 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.107566 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.107667 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.107690 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.107719 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.107746 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.211406 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.211465 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.211482 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.211505 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.211521 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.314909 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.314995 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.315014 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.315040 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.315057 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.418332 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.418405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.418427 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.418456 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.418476 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.522164 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.522209 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.522229 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.522253 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.522272 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.624050 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.624089 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.624099 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.624113 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.624124 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.662660 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.662743 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.662756 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.662771 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.662782 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.683466 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:30Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.686686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.686911 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.686973 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.687040 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.687102 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.704112 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:30Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.707941 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.708074 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.708207 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.708433 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.708606 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.720822 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:30Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.724286 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.724308 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.724317 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.724330 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.724340 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.735255 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:30Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.738516 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.738579 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.738597 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.738654 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.738678 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.750423 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:30Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:30 crc kubenswrapper[5018]: E1014 06:50:30.750694 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.752651 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.752712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.752733 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.752760 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.752778 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.855725 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.855772 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.855793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.855820 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.855841 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.958716 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.958748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.958760 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.958775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:30 crc kubenswrapper[5018]: I1014 06:50:30.958785 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:30Z","lastTransitionTime":"2025-10-14T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.063501 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.063556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.063582 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.063612 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.063689 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.166795 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.166846 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.166862 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.166884 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.166898 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.270615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.270692 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.270704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.270718 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.270729 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376068 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376121 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376141 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376183 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376216 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.376180 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.389941 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.402693 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.415370 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.427285 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.443579 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.459084 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.478215 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.479031 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.479117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.479134 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.479166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.479226 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.495406 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.509119 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.527000 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.541673 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.554662 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.564900 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.582058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.582091 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.582101 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.582118 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.582128 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.586185 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.597988 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.604809 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:31 crc kubenswrapper[5018]: E1014 06:50:31.604909 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.605144 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:31 crc kubenswrapper[5018]: E1014 06:50:31.605195 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.605231 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:31 crc kubenswrapper[5018]: E1014 06:50:31.605276 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.605305 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:31 crc kubenswrapper[5018]: E1014 06:50:31.605338 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.611853 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.622994 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:31Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.684488 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.684796 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.684905 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.685079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.685251 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.789498 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.789545 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.789574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.789597 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.789610 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.892948 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.893200 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.893386 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.893590 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.893717 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.995751 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.995783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.995791 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.995804 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:31 crc kubenswrapper[5018]: I1014 06:50:31.995814 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:31Z","lastTransitionTime":"2025-10-14T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.098670 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.099011 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.099300 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.099519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.099771 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.202059 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.202093 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.202104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.202118 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.202129 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.305348 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.305790 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.305995 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.306197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.306604 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.409842 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.410731 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.411112 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.411332 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.411496 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.514990 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.515030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.515042 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.515058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.515070 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.617443 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.617836 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.617985 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.618126 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.618250 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.720780 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.720998 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.721059 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.721166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.721232 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.825324 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.825668 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.825821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.826007 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.826164 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.928474 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.928526 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.928539 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.928558 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:32 crc kubenswrapper[5018]: I1014 06:50:32.928571 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:32Z","lastTransitionTime":"2025-10-14T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.031002 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.031043 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.031058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.031079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.031090 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.133733 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.133774 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.133783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.133798 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.133809 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.236544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.236578 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.236589 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.236604 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.236615 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.339464 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.339531 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.339551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.339575 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.339593 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.442771 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.442843 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.442866 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.442898 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.442916 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.546311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.546362 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.546371 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.546387 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.546397 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.604148 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.604194 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.604175 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.604166 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:33 crc kubenswrapper[5018]: E1014 06:50:33.604346 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:33 crc kubenswrapper[5018]: E1014 06:50:33.604491 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:33 crc kubenswrapper[5018]: E1014 06:50:33.604699 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:33 crc kubenswrapper[5018]: E1014 06:50:33.604842 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.650043 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.650109 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.650135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.650164 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.650187 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.752422 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.752497 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.752530 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.752555 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.752573 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.856880 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.856936 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.856952 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.856974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.856990 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.961458 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.961525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.961552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.961591 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:33 crc kubenswrapper[5018]: I1014 06:50:33.961661 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:33Z","lastTransitionTime":"2025-10-14T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.070487 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.070560 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.070578 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.070602 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.070648 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.172954 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.173025 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.173048 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.173069 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.173084 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.276247 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.276278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.276289 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.276304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.276317 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.378353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.378390 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.378424 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.378441 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.378453 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.480826 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.480864 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.480875 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.480890 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.480901 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.584602 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.584696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.584712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.584737 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.584753 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.605972 5018 scope.go:117] "RemoveContainer" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.623872 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.643877 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.665421 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.687926 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.688268 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.688448 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.688662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.688867 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.697846 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.713112 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.728080 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.742070 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.757754 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.771125 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.787355 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.791874 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.791925 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.791942 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.791968 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.791982 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.800207 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.810840 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.822415 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.832665 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.846336 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.861000 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.870769 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:34Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.895717 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.895754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.895770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.895972 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.896006 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.999312 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.999345 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.999354 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.999367 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:34 crc kubenswrapper[5018]: I1014 06:50:34.999376 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:34Z","lastTransitionTime":"2025-10-14T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.004704 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/1.log" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.008076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.008531 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.028259 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.044347 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.057502 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.075593 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.101242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.101275 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.101284 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.101296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.101306 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.103606 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.117853 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.136319 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.152098 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.163951 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.183189 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.194741 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.203884 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.203913 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.203921 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.203936 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.203947 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.207884 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.220454 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.239759 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.254916 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.278172 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.300526 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:35Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.306292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.306340 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.306352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.306368 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.306384 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.408657 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.408693 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.408704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.408718 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.408727 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.433293 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.433537 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.433613 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433659 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433675 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:51:07.433642504 +0000 UTC m=+84.017689141 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433729 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433744 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:50:51.433723196 +0000 UTC m=+68.017769843 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433786 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:51:07.433772968 +0000 UTC m=+84.017819705 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.433779 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433833 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.433874 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:51:07.43386504 +0000 UTC m=+84.017911677 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.510866 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.510900 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.510908 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.510922 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.510931 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.534903 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.534978 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535270 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535297 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535307 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535363 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:51:07.535345698 +0000 UTC m=+84.119392325 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535361 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535397 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535411 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.535476 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:51:07.535457541 +0000 UTC m=+84.119504178 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.604705 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.604745 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.604777 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.604707 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.604831 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.604921 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.605106 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:35 crc kubenswrapper[5018]: E1014 06:50:35.605230 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.614079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.614109 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.614118 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.614131 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.614142 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.716716 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.716752 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.716761 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.716775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.716785 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.820510 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.820580 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.820601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.820659 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.820683 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.923863 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.923922 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.923945 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.923976 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:35 crc kubenswrapper[5018]: I1014 06:50:35.924005 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:35Z","lastTransitionTime":"2025-10-14T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.015231 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/2.log" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.016081 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/1.log" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.019255 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" exitCode=1 Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.019303 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.019352 5018 scope.go:117] "RemoveContainer" containerID="ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.020310 5018 scope.go:117] "RemoveContainer" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" Oct 14 06:50:36 crc kubenswrapper[5018]: E1014 06:50:36.021058 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.029324 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.029390 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.029420 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.029688 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.029712 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.045872 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.066062 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.084966 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.107278 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea97ba931b85ad85ab240b57deed5aee303c29e9bc84e9619bf9c2b08fb1f459\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:17Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 06:50:16.993347 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 06:50:16.993377 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 06:50:16.993418 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:16.993477 6453 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:16.993494 6453 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:16.993505 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:16.993529 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:16.993544 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:16.993568 6453 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:16.993663 6453 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:16.994218 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:16.994286 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 06:50:16.994319 6453 factory.go:656] Stopping watch factory\\\\nI1014 06:50:16.994341 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:16.994386 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.123172 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.134401 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.134436 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.134446 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.134462 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.134474 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.139857 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.157447 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.181404 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.197510 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.214763 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.230110 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.237361 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.237420 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.237439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.237465 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.237483 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.245293 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.256449 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.268445 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.277991 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.288357 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.303832 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:36Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.340594 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.340642 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.340653 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.340670 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.340681 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.443583 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.443764 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.443788 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.443849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.443868 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.547062 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.547159 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.547180 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.547208 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.547227 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.650425 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.650507 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.650527 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.650559 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.650578 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.753697 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.753749 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.753767 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.753790 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.753806 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.856769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.856806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.856817 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.856833 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.856845 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.960069 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.960152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.960216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.960248 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:36 crc kubenswrapper[5018]: I1014 06:50:36.960269 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:36Z","lastTransitionTime":"2025-10-14T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.025653 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/2.log" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.032289 5018 scope.go:117] "RemoveContainer" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" Oct 14 06:50:37 crc kubenswrapper[5018]: E1014 06:50:37.032492 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.048002 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.063687 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.063748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.063768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.063795 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.063814 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.066597 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.090379 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.106233 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.124671 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.142747 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.166056 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.167361 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.167404 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.167418 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.167445 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.167459 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.193664 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.212723 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.232025 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.254616 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.270315 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.270370 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.270394 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.270457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.270508 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.279076 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.302048 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.320968 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.340022 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.355900 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.374297 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.374344 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.374364 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.374388 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.374407 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.376475 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:37Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.479465 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.479537 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.479553 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.479578 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.479595 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.582902 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.583014 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.583042 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.583073 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.583096 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.604721 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.604759 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.604806 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.604721 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:37 crc kubenswrapper[5018]: E1014 06:50:37.604919 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:37 crc kubenswrapper[5018]: E1014 06:50:37.605041 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:37 crc kubenswrapper[5018]: E1014 06:50:37.605143 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:37 crc kubenswrapper[5018]: E1014 06:50:37.605252 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.687212 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.687322 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.687342 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.687368 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.687388 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.791660 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.791749 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.791769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.791802 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.791823 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.894769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.894828 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.894845 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.894868 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.894885 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.998567 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.998665 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.998685 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.998712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:37 crc kubenswrapper[5018]: I1014 06:50:37.998730 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:37Z","lastTransitionTime":"2025-10-14T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.101920 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.102009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.102029 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.102056 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.102083 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.206074 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.206132 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.206152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.206184 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.206208 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.309496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.309671 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.309696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.309723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.309740 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.412802 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.412862 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.412879 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.412905 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.412923 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.516980 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.517136 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.517163 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.517193 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.517213 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.620033 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.620099 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.620116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.620142 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.620163 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.722671 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.722709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.722739 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.722752 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.722761 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.830865 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.830936 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.830961 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.830992 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.831015 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.934413 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.934480 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.934525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.934550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:38 crc kubenswrapper[5018]: I1014 06:50:38.934567 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:38Z","lastTransitionTime":"2025-10-14T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.037963 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.038027 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.038047 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.038076 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.038096 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.141599 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.141699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.141718 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.141745 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.141763 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.245910 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.245983 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.246003 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.246027 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.246046 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.349479 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.349572 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.349597 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.349657 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.349680 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.453006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.453070 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.453087 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.453111 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.453130 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.557204 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.557263 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.557284 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.557308 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.557326 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.604962 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:39 crc kubenswrapper[5018]: E1014 06:50:39.605139 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.605418 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:39 crc kubenswrapper[5018]: E1014 06:50:39.605525 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.605824 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.605936 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:39 crc kubenswrapper[5018]: E1014 06:50:39.606137 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:39 crc kubenswrapper[5018]: E1014 06:50:39.606257 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.660360 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.660416 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.660434 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.660458 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.660476 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.763396 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.763464 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.763481 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.763507 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.763524 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.866989 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.867050 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.867071 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.867097 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.867114 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.970775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.970828 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.970846 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.970870 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:39 crc kubenswrapper[5018]: I1014 06:50:39.970890 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:39Z","lastTransitionTime":"2025-10-14T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.074313 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.074496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.074545 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.074573 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.074591 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.178328 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.178379 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.178397 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.178417 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.178430 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.281998 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.282160 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.282183 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.282207 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.282224 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.385318 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.385375 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.385393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.385419 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.385437 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.487973 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.488028 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.488045 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.488070 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.488086 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.591581 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.591691 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.591719 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.591749 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.591773 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.694729 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.694790 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.694809 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.694833 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.694851 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.797655 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.797716 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.797761 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.797789 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.797804 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.900814 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.900864 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.900875 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.900893 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.900907 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.969652 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.969717 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.969735 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.969763 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:40 crc kubenswrapper[5018]: I1014 06:50:40.969782 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:40Z","lastTransitionTime":"2025-10-14T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:40 crc kubenswrapper[5018]: E1014 06:50:40.989840 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:40Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.000983 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.001094 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.001117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.001145 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.001165 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.021363 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:41Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.026656 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.026738 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.026767 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.026819 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.026843 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.049118 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:41Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.054938 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.055041 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.055082 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.055120 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.055219 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.077743 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:41Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.082340 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.082390 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.082439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.082461 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.082480 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.102160 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:41Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.102391 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.104347 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.104423 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.104445 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.104469 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.104521 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.208307 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.208376 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.208395 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.208423 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.208442 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.311552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.311614 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.311660 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.311686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.311705 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.414750 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.414846 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.414864 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.414889 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.414909 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.518010 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.518071 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.518091 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.518117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.518134 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.604943 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.604985 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.604985 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.605144 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.605289 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.605374 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.605408 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:41 crc kubenswrapper[5018]: E1014 06:50:41.605540 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.621818 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.621887 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.621911 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.621940 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.621959 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.725544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.725612 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.725669 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.725694 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.725711 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.829079 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.829155 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.829173 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.829196 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.829211 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.931850 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.931944 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.931961 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.931985 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:41 crc kubenswrapper[5018]: I1014 06:50:41.932003 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:41Z","lastTransitionTime":"2025-10-14T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.034997 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.035037 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.035050 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.035065 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.035075 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.137538 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.137609 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.137662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.137699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.137726 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.240646 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.240800 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.240825 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.240853 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.240870 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.344185 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.344252 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.344269 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.344292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.344310 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.448160 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.448258 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.448287 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.448320 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.448346 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.551442 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.551520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.551540 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.551564 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.551582 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.654395 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.654452 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.654469 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.654492 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.654509 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.757702 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.757762 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.757779 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.757802 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.757819 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.861746 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.861793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.861808 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.861827 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.861839 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.965230 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.965285 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.965296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.965313 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:42 crc kubenswrapper[5018]: I1014 06:50:42.965326 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:42Z","lastTransitionTime":"2025-10-14T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.067765 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.067824 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.067843 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.067870 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.067887 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.171582 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.171671 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.171693 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.171719 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.171740 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.274882 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.274930 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.274955 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.274979 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.274995 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.377883 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.377948 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.377969 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.377994 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.378212 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.481439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.481506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.481515 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.481532 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.481541 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.584479 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.584574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.584595 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.584674 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.584710 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.604423 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:43 crc kubenswrapper[5018]: E1014 06:50:43.604595 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.604892 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.604978 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:43 crc kubenswrapper[5018]: E1014 06:50:43.605073 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.605135 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:43 crc kubenswrapper[5018]: E1014 06:50:43.605260 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:43 crc kubenswrapper[5018]: E1014 06:50:43.605359 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.688241 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.688297 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.688316 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.688340 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.688357 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.791414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.791474 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.791495 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.791520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.791540 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.894946 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.895016 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.895036 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.895060 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.895080 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.997987 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.998031 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.998042 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.998060 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:43 crc kubenswrapper[5018]: I1014 06:50:43.998073 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:43Z","lastTransitionTime":"2025-10-14T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.100668 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.100724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.100736 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.100754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.100769 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.203521 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.203598 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.203613 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.203657 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.203669 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.307166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.307229 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.307248 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.307273 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.307291 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.410689 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.410768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.410791 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.410839 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.410865 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.513543 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.513594 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.513610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.513701 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.513730 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.617425 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.617487 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.617507 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.617535 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.617558 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.627664 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.649719 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.666168 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.685097 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.707145 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.720135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.720200 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.720218 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.720241 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.720259 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.724201 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.762143 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.781151 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.800497 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.814296 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.823360 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.823507 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.823526 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.823561 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.823588 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.831519 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.854761 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.873228 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.897218 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.918881 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.929699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.929755 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.929776 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.929811 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.929832 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:44Z","lastTransitionTime":"2025-10-14T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.953923 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:44 crc kubenswrapper[5018]: I1014 06:50:44.976977 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:44Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.034370 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.034442 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.034460 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.034486 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.034505 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.137749 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.137818 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.137835 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.137864 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.137888 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.240984 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.241044 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.241064 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.241088 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.241107 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.344526 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.344598 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.344655 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.344693 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.344715 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.447440 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.447500 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.447525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.447552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.447568 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.551346 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.551407 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.551427 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.551453 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.551473 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.604443 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:45 crc kubenswrapper[5018]: E1014 06:50:45.604611 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.604740 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:45 crc kubenswrapper[5018]: E1014 06:50:45.605351 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.605428 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:45 crc kubenswrapper[5018]: E1014 06:50:45.605577 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.605450 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:45 crc kubenswrapper[5018]: E1014 06:50:45.605936 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.655175 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.655239 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.655264 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.655294 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.655315 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.758496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.758558 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.758575 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.758601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.758651 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.862770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.862827 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.862844 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.862869 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.862886 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.966006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.966062 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.966078 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.966103 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:45 crc kubenswrapper[5018]: I1014 06:50:45.966120 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:45Z","lastTransitionTime":"2025-10-14T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.068640 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.068683 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.068696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.068712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.068724 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.171365 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.171405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.171416 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.171433 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.171445 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.274196 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.274265 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.274282 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.274308 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.274326 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.377692 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.377748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.377765 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.377789 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.377825 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.480471 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.480817 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.480916 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.481033 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.481131 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.584223 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.584280 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.584298 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.584322 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.584337 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.687030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.687091 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.687112 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.687137 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.687155 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.789692 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.799895 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.800145 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.800320 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.800466 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.908137 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.908200 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.908221 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.908259 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:46 crc kubenswrapper[5018]: I1014 06:50:46.908280 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:46Z","lastTransitionTime":"2025-10-14T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.012052 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.012104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.012122 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.012146 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.012164 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.115599 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.115686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.115704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.115729 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.115747 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.218722 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.218788 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.218806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.218833 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.218851 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.322006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.322058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.322076 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.322098 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.322114 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.424753 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.424821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.424844 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.424874 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.424898 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.527586 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.527745 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.527768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.527790 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.527809 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.604798 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.604828 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.604807 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.604976 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:47 crc kubenswrapper[5018]: E1014 06:50:47.605131 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:47 crc kubenswrapper[5018]: E1014 06:50:47.605273 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:47 crc kubenswrapper[5018]: E1014 06:50:47.605442 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:47 crc kubenswrapper[5018]: E1014 06:50:47.605609 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.630442 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.630502 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.630525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.630556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.630579 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.733371 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.733432 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.733448 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.733472 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.733489 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.836409 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.836496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.836522 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.836554 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.836575 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.940394 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.940436 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.940447 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.940465 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:47 crc kubenswrapper[5018]: I1014 06:50:47.940477 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:47Z","lastTransitionTime":"2025-10-14T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.044191 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.044260 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.044278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.044304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.044322 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.147144 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.147216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.147236 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.147262 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.147279 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.250350 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.250394 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.250412 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.250435 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.250452 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.353143 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.353209 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.353228 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.353252 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.353269 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.456892 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.457358 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.457380 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.457405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.457423 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.559843 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.559943 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.559990 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.560020 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.560037 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.663647 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.663700 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.663711 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.663728 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.663740 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.765436 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.765472 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.765483 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.765499 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.765522 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.868419 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.868496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.868514 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.868536 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.868553 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.970556 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.970599 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.970651 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.970684 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:48 crc kubenswrapper[5018]: I1014 06:50:48.970701 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:48Z","lastTransitionTime":"2025-10-14T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.073576 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.073661 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.073687 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.073709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.073725 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.176251 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.176310 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.176324 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.176343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.176356 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.279959 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.280016 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.280029 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.280051 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.280064 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.383768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.383831 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.383845 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.383872 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.383889 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.487356 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.487775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.487931 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.488076 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.488209 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.591611 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.591715 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.591744 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.591779 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.591818 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.604387 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.604439 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.604523 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:49 crc kubenswrapper[5018]: E1014 06:50:49.604685 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.604821 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:49 crc kubenswrapper[5018]: E1014 06:50:49.604982 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:49 crc kubenswrapper[5018]: E1014 06:50:49.605605 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:49 crc kubenswrapper[5018]: E1014 06:50:49.605705 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.606232 5018 scope.go:117] "RemoveContainer" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" Oct 14 06:50:49 crc kubenswrapper[5018]: E1014 06:50:49.606728 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.693959 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.694010 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.694019 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.694034 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.694046 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.796664 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.796759 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.796778 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.797216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.797395 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.899345 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.899377 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.899388 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.899404 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:49 crc kubenswrapper[5018]: I1014 06:50:49.899415 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:49Z","lastTransitionTime":"2025-10-14T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.001435 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.001467 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.001475 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.001489 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.001499 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.103669 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.103707 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.103717 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.103730 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.103741 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.205987 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.206058 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.206082 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.206113 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.206135 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.309463 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.309531 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.309554 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.309585 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.309605 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.412404 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.412458 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.412476 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.412498 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.412516 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.515226 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.515265 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.515282 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.515299 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.515308 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.617773 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.617829 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.617846 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.617867 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.617885 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.720676 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.720714 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.720723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.720738 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.720749 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.823404 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.823478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.823506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.823537 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.823561 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.926044 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.926085 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.926104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.926129 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:50 crc kubenswrapper[5018]: I1014 06:50:50.926146 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:50Z","lastTransitionTime":"2025-10-14T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.027958 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.027993 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.028004 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.028020 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.028034 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.130466 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.130510 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.130519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.130535 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.130545 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.234138 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.234191 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.234213 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.234241 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.234264 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.336551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.336584 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.336592 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.336604 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.336630 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.389249 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.389281 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.389291 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.389311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.389320 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.404054 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:51Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.408327 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.408448 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.408514 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.408550 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.408607 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.424004 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:51Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.427923 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.427965 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.427974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.427990 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.427999 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.444755 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:51Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.448343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.448399 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.448410 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.448429 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.448441 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.464649 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:51Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.468039 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.468083 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.468096 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.468117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.468131 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.482720 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:51Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.482960 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.485285 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.485336 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.485353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.485380 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.485398 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.523337 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.523473 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.523540 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:51:23.523522907 +0000 UTC m=+100.107569554 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.590292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.590327 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.590336 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.590352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.590362 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.604786 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.604813 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.604792 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.604889 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.604791 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.605017 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.605062 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:51 crc kubenswrapper[5018]: E1014 06:50:51.605183 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.693030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.693067 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.693077 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.693097 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.693108 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.795939 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.795990 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.796001 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.796019 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.796031 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.898590 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.898681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.898698 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.898723 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:51 crc kubenswrapper[5018]: I1014 06:50:51.898739 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:51Z","lastTransitionTime":"2025-10-14T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.000855 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.000896 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.000908 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.000924 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.000935 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.103246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.103316 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.103338 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.103393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.103414 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.205459 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.205497 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.205506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.205521 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.205531 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.308246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.308320 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.308343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.308372 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.308397 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.411082 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.411135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.411148 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.411169 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.411181 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.513748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.513804 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.513823 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.513848 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.513866 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.616983 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.617060 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.617083 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.617112 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.617135 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.720597 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.720702 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.720721 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.720748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.720767 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.823574 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.823612 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.823634 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.823648 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.823658 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.926257 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.926295 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.926303 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.926320 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:52 crc kubenswrapper[5018]: I1014 06:50:52.926329 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:52Z","lastTransitionTime":"2025-10-14T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.031533 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.031607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.031662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.031689 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.031707 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.087673 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/0.log" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.087765 5018 generic.go:334] "Generic (PLEG): container finished" podID="f1aa622b-96ac-42aa-8bd1-8fa174e65602" containerID="e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb" exitCode=1 Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.087834 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerDied","Data":"e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.089015 5018 scope.go:117] "RemoveContainer" containerID="e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.105554 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.124938 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.134520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.134586 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.134605 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.134785 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.134905 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.148128 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.166499 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.178978 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.190447 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.203244 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.216552 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.230427 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.237586 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.237680 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.237699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.237725 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.237742 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.255797 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.273106 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.285297 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.296927 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.312085 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.325600 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.340457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.340499 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.340508 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.340522 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.340534 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.346734 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.359664 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:53Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.442493 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.442534 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.442547 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.442563 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.442575 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.545658 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.545698 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.545710 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.545726 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.545738 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.604521 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.604579 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.604724 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:53 crc kubenswrapper[5018]: E1014 06:50:53.604853 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.604890 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:53 crc kubenswrapper[5018]: E1014 06:50:53.605015 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:53 crc kubenswrapper[5018]: E1014 06:50:53.605045 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:53 crc kubenswrapper[5018]: E1014 06:50:53.605104 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.648209 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.648247 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.648256 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.648270 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.648280 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.751466 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.751520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.751538 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.751560 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.751577 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.854359 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.854402 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.854414 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.854431 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.854444 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.957436 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.957500 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.957519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.957544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:53 crc kubenswrapper[5018]: I1014 06:50:53.957562 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:53Z","lastTransitionTime":"2025-10-14T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.060142 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.060186 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.060197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.060215 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.060227 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.093035 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/0.log" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.093118 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerStarted","Data":"7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.112073 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.126571 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.142928 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.162449 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.162491 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.162506 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.162527 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.162542 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.168334 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.183192 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.197149 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.209044 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.226344 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.237649 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.249926 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.261519 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.265356 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.265433 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.265446 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.265461 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.265474 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.271813 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.287010 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.302597 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.312843 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.326967 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.340103 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.368130 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.368158 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.368171 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.368186 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.368197 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.471217 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.471278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.471296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.471322 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.471338 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.573728 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.573765 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.573776 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.573792 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.573803 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.618239 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.631104 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.648436 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.670813 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.676512 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.676594 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.676607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.676646 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.676659 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.683454 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.698144 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.709687 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.731460 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.745152 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.761953 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.776887 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.779166 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.779372 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.779430 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.779454 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.779475 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.789501 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.807464 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.823078 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.838643 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.856107 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.868773 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:50:54Z is after 2025-08-24T17:21:41Z" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.881937 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.882002 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.882014 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.882057 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.882071 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.988312 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.988371 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.988386 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.988405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:54 crc kubenswrapper[5018]: I1014 06:50:54.988420 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:54Z","lastTransitionTime":"2025-10-14T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.090679 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.090714 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.090725 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.090743 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.090756 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.192793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.192848 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.192865 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.192890 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.192908 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.295475 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.295551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.295573 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.295602 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.295655 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.398190 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.398480 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.398569 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.398687 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.398779 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.501406 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.501663 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.501757 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.501831 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.501896 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604160 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604188 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604211 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604298 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604362 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604389 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604399 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604411 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.604421 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: E1014 06:50:55.604468 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:55 crc kubenswrapper[5018]: E1014 06:50:55.604718 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:55 crc kubenswrapper[5018]: E1014 06:50:55.604837 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:55 crc kubenswrapper[5018]: E1014 06:50:55.604994 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.707329 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.707395 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.707413 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.707437 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.707456 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.809533 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.809583 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.809592 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.809608 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.809644 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.912672 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.912746 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.912770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.912794 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:55 crc kubenswrapper[5018]: I1014 06:50:55.912812 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:55Z","lastTransitionTime":"2025-10-14T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.015612 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.015692 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.015709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.015733 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.015750 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.117394 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.117457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.117477 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.117501 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.117516 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.219923 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.219997 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.220020 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.220050 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.220072 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.323505 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.323577 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.323598 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.323673 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.323693 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.427168 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.427222 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.427234 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.427254 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.427267 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.531995 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.532057 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.532075 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.532101 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.532120 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.634339 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.634386 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.634401 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.634423 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.634439 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.737557 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.737610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.737638 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.737656 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.737669 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.839566 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.839607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.839645 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.839662 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.839673 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.942259 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.942300 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.942310 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.942325 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:56 crc kubenswrapper[5018]: I1014 06:50:56.942337 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:56Z","lastTransitionTime":"2025-10-14T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.045575 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.045655 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.045672 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.045699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.045719 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.148148 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.148292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.148359 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.148386 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.148403 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.251969 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.252022 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.252039 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.252063 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.252081 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.355566 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.355727 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.355748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.355772 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.355790 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.458736 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.458799 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.458821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.458847 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.458865 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.561696 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.561738 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.561750 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.561766 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.561778 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.604670 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.604721 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.604670 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:57 crc kubenswrapper[5018]: E1014 06:50:57.604773 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.604954 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:57 crc kubenswrapper[5018]: E1014 06:50:57.604965 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:57 crc kubenswrapper[5018]: E1014 06:50:57.605007 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:57 crc kubenswrapper[5018]: E1014 06:50:57.605074 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.663776 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.663849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.663867 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.663895 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.663913 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.767116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.767156 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.767169 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.767186 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.767196 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.870734 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.870769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.870779 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.870793 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.870802 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.973949 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.974006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.974023 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.974047 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:57 crc kubenswrapper[5018]: I1014 06:50:57.974065 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:57Z","lastTransitionTime":"2025-10-14T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.077046 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.077088 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.077099 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.077115 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.077126 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.179439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.179474 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.179482 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.179496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.179506 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.282006 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.282091 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.282119 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.282155 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.282182 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.385562 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.385610 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.385639 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.385656 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.385668 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.488669 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.488704 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.488717 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.488731 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.488740 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.591114 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.591159 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.591172 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.591189 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.591201 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.693987 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.694047 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.694064 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.694095 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.694113 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.797544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.797665 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.797731 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.797764 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.797785 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.900911 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.900967 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.900989 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.901015 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:58 crc kubenswrapper[5018]: I1014 06:50:58.901036 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:58Z","lastTransitionTime":"2025-10-14T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.003682 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.004009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.004037 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.004068 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.004092 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.107242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.107288 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.107311 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.107339 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.107361 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.210504 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.210564 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.210584 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.210615 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.210675 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.313214 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.313266 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.313278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.313296 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.313309 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.416276 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.416329 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.416340 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.416358 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.416374 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.519761 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.519838 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.519860 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.519888 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.519911 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.604272 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.604316 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.604370 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.604469 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:50:59 crc kubenswrapper[5018]: E1014 06:50:59.604708 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:50:59 crc kubenswrapper[5018]: E1014 06:50:59.604904 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:50:59 crc kubenswrapper[5018]: E1014 06:50:59.605047 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:50:59 crc kubenswrapper[5018]: E1014 06:50:59.605122 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.623287 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.623360 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.623416 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.623448 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.623471 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.726264 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.726325 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.726344 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.726370 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.726387 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.829135 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.829187 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.829205 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.829229 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.829244 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.932083 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.932154 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.932176 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.932203 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:50:59 crc kubenswrapper[5018]: I1014 06:50:59.932220 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:50:59Z","lastTransitionTime":"2025-10-14T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.034480 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.034553 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.034579 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.034609 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.034666 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.137783 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.137853 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.137873 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.137895 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.137913 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.241371 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.241426 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.241443 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.241467 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.241486 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.345086 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.345142 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.345161 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.345189 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.345211 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.448481 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.448545 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.448565 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.448590 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.448609 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.551819 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.551904 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.551929 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.551959 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.551985 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.655469 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.655525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.655544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.655568 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.655586 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.758205 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.758238 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.758246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.758261 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.758271 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.861473 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.861512 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.861524 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.861540 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.861551 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.967893 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.967998 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.968038 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.968077 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:00 crc kubenswrapper[5018]: I1014 06:51:00.968106 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:00Z","lastTransitionTime":"2025-10-14T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.070758 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.070811 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.070828 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.070853 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.070871 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.173588 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.173686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.173711 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.173745 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.173767 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.276072 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.276152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.276175 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.276200 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.276219 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.379131 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.379197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.379220 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.379252 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.379271 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.482599 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.482702 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.482725 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.482760 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.482779 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.586568 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.586667 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.586686 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.586713 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.586731 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.604213 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.604283 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.604226 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.604222 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.604524 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.604753 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.604928 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.605081 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.619210 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.619264 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.619283 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.619304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.619323 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.640392 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:01Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.646333 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.646375 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.646392 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.646415 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.646433 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.666710 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:01Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.671553 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.671601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.671642 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.671666 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.671689 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.691650 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:01Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.696836 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.696903 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.696931 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.696957 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.696971 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.715865 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:01Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.720849 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.720899 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.720918 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.720945 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.720964 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.741512 5018 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T06:51:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"772668f9-2246-4743-84c4-813da958e906\\\",\\\"systemUUID\\\":\\\"ca6145eb-67b2-4317-a65f-9b8f914c2ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:01Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:01 crc kubenswrapper[5018]: E1014 06:51:01.741760 5018 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.743712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.743774 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.743798 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.743827 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.743848 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.846917 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.846982 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.847004 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.847033 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.847052 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.949937 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.950009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.950036 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.950068 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:01 crc kubenswrapper[5018]: I1014 06:51:01.950090 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:01Z","lastTransitionTime":"2025-10-14T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.052752 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.052808 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.052825 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.052847 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.052864 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.155102 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.155157 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.155175 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.155199 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.155217 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.264688 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.264742 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.264771 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.264800 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.264822 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.368055 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.368109 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.368131 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.368159 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.368179 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.471764 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.471822 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.471842 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.471872 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.471895 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.575221 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.575294 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.575321 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.575349 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.575369 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.605610 5018 scope.go:117] "RemoveContainer" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.677238 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.677292 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.677304 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.677325 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.677338 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.779678 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.779712 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.779722 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.779735 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.779743 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.883256 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.883293 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.883303 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.883317 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.883326 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.986148 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.986204 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.986222 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.986246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:02 crc kubenswrapper[5018]: I1014 06:51:02.986264 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:02Z","lastTransitionTime":"2025-10-14T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.089472 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.089551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.089580 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.089613 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.089672 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.125455 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/2.log" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.130305 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.130990 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.158455 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.178184 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.192334 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.192382 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.192395 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.192413 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.192427 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.193727 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.215400 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:51:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.232426 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.247300 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.260678 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.276506 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.288525 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.295376 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.295419 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.295435 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.295457 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.295472 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.307244 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.318858 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.333689 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.346715 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.356842 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.367680 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.379809 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.391543 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.397952 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.397974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.397985 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.397998 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.398007 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.500974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.501009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.501018 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.501032 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.501041 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604033 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604075 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604093 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604116 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604133 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604760 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604809 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604885 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:03 crc kubenswrapper[5018]: E1014 06:51:03.604919 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.604765 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:03 crc kubenswrapper[5018]: E1014 06:51:03.605014 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:03 crc kubenswrapper[5018]: E1014 06:51:03.605183 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:03 crc kubenswrapper[5018]: E1014 06:51:03.605270 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.707213 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.707278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.707301 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.707331 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.707353 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.810525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.810601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.810681 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.810715 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.810737 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.914108 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.914157 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.914175 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.914199 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:03 crc kubenswrapper[5018]: I1014 06:51:03.914216 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:03Z","lastTransitionTime":"2025-10-14T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.017715 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.017775 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.017792 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.017816 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.017835 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.120670 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.120745 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.120756 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.120770 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.120780 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.137419 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/3.log" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.138507 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/2.log" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.142099 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" exitCode=1 Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.142161 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.142211 5018 scope.go:117] "RemoveContainer" containerID="dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.143424 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 06:51:04 crc kubenswrapper[5018]: E1014 06:51:04.143719 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.174913 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.194666 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.210314 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.224115 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.224173 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.224197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.224226 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.224246 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.233465 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.252135 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.269283 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.282738 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.299373 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.313683 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.327843 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.327882 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.327894 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.327911 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.327923 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.328434 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.344790 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.355812 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.369141 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.387144 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.401379 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.418005 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.436834 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.436910 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.436936 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.436968 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.436994 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.441228 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:51:03Z\\\",\\\"message\\\":\\\"026 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1014 06:51:03.543239 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1014 06:51:03.543246 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt\\\\nI1014 06:51:03.543249 7026 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1014 06:51:03.543239 7026 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z]\\\\nI1014 06:51\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:51:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.540132 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.540193 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.540216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.540244 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.540267 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.623974 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.639113 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.644551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.644607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.644633 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.644665 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.644680 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.659584 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.692129 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.711005 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.725893 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.744452 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.751394 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.751611 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.751666 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.751701 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.751727 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.769541 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.782764 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.796606 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.813442 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.835402 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dee6255fafa557e621a2db8e2407d6a0b7a867df9dd5b799836e804234ed0896\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:35Z\\\",\\\"message\\\":\\\"er.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 06:50:35.532311 6665 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 06:50:35.532347 6665 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 06:50:35.532356 6665 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 06:50:35.532411 6665 factory.go:656] Stopping watch factory\\\\nI1014 06:50:35.532429 6665 ovnkube.go:599] Stopped ovnkube\\\\nI1014 06:50:35.532462 6665 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 06:50:35.532475 6665 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 06:50:35.532483 6665 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 06:50:35.532491 6665 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 06:50:35.532498 6665 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 06:50:35.532506 6665 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 06:50:35.532513 6665 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1014 06:50:35.532521 6665 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 06:50:35.532529 6665 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 06:50:35.532540 6665 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1014 06:50:35.532608 6665 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:51:03Z\\\",\\\"message\\\":\\\"026 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1014 06:51:03.543239 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1014 06:51:03.543246 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt\\\\nI1014 06:51:03.543249 7026 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1014 06:51:03.543239 7026 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z]\\\\nI1014 06:51\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:51:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.851905 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.855476 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.855508 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.855519 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.855537 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.855549 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.865817 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.880348 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.895148 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.914995 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:04Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.959117 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.959204 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.959238 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.959270 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:04 crc kubenswrapper[5018]: I1014 06:51:04.959289 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:04Z","lastTransitionTime":"2025-10-14T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.062176 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.062279 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.062314 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.062352 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.062368 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.148436 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/3.log" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.154396 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 06:51:05 crc kubenswrapper[5018]: E1014 06:51:05.154764 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.165437 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.165509 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.165533 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.165562 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.165584 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.177360 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c78f6c2-0123-4199-82c4-b8df0d48e381\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://160a5c284a57ae43a9137460ea214cbd16be7d68649e4eb3ca6880a4b9aeb2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4328544ca9333311f1f4867b94eb9a6de75376c3e04100ead994e7d7bafda95a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29021b21cf98c843093c442eed9eb48f6b3cc80f17c0715449a709754f557cb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8146e7e55441b6ccddb145769bbca66dd5855b2585997d5d32b8912c4b0236\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://964a4e6f5fb9cf80396ec607f796de6006248a8603c6b017fcfafa3d83caf100\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1014 06:50:03.459952 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1014 06:50:03.459975 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1014 06:50:03.460151 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1014 06:50:03.460181 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1014 06:50:03.460364 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1760424587\\\\\\\\\\\\\\\" (2025-10-14 06:49:47 +0000 UTC to 2025-11-13 06:49:48 +0000 UTC (now=2025-10-14 06:50:03.460310372 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1760424588\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1760424588\\\\\\\\\\\\\\\" (2025-10-14 05:49:48 +0000 UTC to 2026-10-14 05:49:48 +0000 UTC (now=2025-10-14 06:50:03.460703964 +0000 UTC))\\\\\\\"\\\\nI1014 06:50:03.460760 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2552210337/tls.crt::/tmp/serving-cert-2552210337/tls.key\\\\\\\"\\\\nI1014 06:50:03.460816 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1014 06:50:03.460854 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1014 06:50:03.460891 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1014 06:50:03.461766 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8482d04e122ffd7481991dbb400f228652e69b327beef22bb200393dd31ddbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6bcfc24e24ceecc770a1964e331cc9e2856e8933b74219fd90ad8be0bfb5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.196483 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56ee799a192e2215eec7f77aead7dc581f89e45102be091539ca19e97adeffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.216378 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r7ftx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1aa622b-96ac-42aa-8bd1-8fa174e65602\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:50:52Z\\\",\\\"message\\\":\\\"2025-10-14T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc\\\\n2025-10-14T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7e8cdac3-66d5-4b9b-bc0e-2051234888bc to /host/opt/cni/bin/\\\\n2025-10-14T06:50:07Z [verbose] multus-daemon started\\\\n2025-10-14T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-10-14T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhkkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r7ftx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.231982 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2530b46d-5f3a-4495-af23-3ad85327c99c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xwjj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gl9fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.255430 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e07068-0c9a-4200-b507-e98dca87d535\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b384013da2863b8dde541a13edc51d099ecffdc82f1a0abf8fe754c63473752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a71e0305d55c1389ebc09940bd8cea2e1ae867d4a5079fcbbb382b50adf8a31\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c68218f6c80af566c470ff9e56dcc9d5a4907809588c67a30da97ecc1496524\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e02b473fd0a0fa61cd2e5ce0ba37f2ee7262c751aee1e1ed1c7e099e1f9a1b68\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7518fbbb76c694c5b95e03b042f47fbb47aeeabda552d6bbd21ff02bd8eced77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54be4007dccf96431e6649ab3506b2faad148d63860a2e28e40447af685a2342\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca3955d59ccbaa771cb83cee20cbba981d983c9d3803587930986ae169748cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj8n2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fzdc9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.268988 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.269353 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.269721 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.269932 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.270076 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.270277 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"898b50bd-ba6a-481e-8337-11aaa1d75199\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61daf158b2d6f12c484421bd6d46dcf2bf63d73ca31e296d7307df00cfd910d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a32558245d7b65dd0f6faa12219205b9546fb33beffdfc980be34707d55a73e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dcdjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.284061 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c91017e7-e1c7-4151-9104-742d3be6520b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23ef514febdfe7f956924624a5249ac09290cfb21a89f060829bab09cc5a9f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bf68ebfc5f1769a9f92dc3667c0dd6a06fe45d1ad1d516f0f73c6cdc7e7bb1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6adc3a4323b07fc0ee6b80d586d59c0f03383717acd4401d4aba859cd11c6bbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e995e6d9598315ea320f2d41d183ba30477fb3e3189a1656a2b53ead70e6df5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:49:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.303493 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.318189 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-db8xr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d6c48b-2053-47ab-b68c-28442219d457\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58cd75ae8c5d5fb49481ac8b5061cb39f386c8427f6a7830db8f24d4a1f26fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ksfw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-db8xr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.334448 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"124fa7dd-4f8d-4d66-9223-f6bcc6df7242\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33073e7a70a541d9b152d31c5ee1270cf51f29079453cdb094fed797d8855e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-scgt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-g7t5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.349422 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17006a80-394f-4518-98c0-469d3579836b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:49:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e965f8e356872350cfb985f0b32fa0a008a76556bc7a543b6a46b0bb77e5007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc9cafa682daabf8ac6cd069a3df45fa7091880dd4b6a0ea8b4f63223f0b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9689c4c2a728c4320c9c45b3097bef4b8d969fd231e0d8613509101454978741\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6d411588f5167002618853312c60883e88e61e617baf574b2851ccec187e72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:49:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.365530 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.373133 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.373233 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.373253 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.373278 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.373298 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.388111 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f328f33ea54bb1d0a4a218dcab113bb6e970973231066a13a0665efdc71ac1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5489ec181aeb24f39cf19521ddca28eef6bc9121cb261e4e4f22b3cc5285d39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.419456 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99edde9a-4116-4bdf-84ff-863e9e6ed55e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T06:51:03Z\\\",\\\"message\\\":\\\"026 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1014 06:51:03.543239 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1014 06:51:03.543246 7026 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt\\\\nI1014 06:51:03.543249 7026 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1014 06:51:03.543239 7026 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:03Z is after 2025-08-24T17:21:41Z]\\\\nI1014 06:51\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T06:51:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn5tp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c2zpf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.441381 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b3dace9d24a504b83da715c3f60529f6b878dd5130e01fd685bfa304d3b929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.461972 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.475021 5018 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pwgtv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"696c1ace-2829-43a4-9103-a91c85ed06ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f749086979c289aab921b69271dc813956821e7b3f5cb512d3f57b9c2da941c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xlwr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T06:50:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pwgtv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T06:51:05Z is after 2025-08-24T17:21:41Z" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.477009 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.477078 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.477105 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.477138 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.477164 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.581439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.581489 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.581515 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.581540 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.581553 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.604836 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.604848 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.604906 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.604848 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:05 crc kubenswrapper[5018]: E1014 06:51:05.604961 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:05 crc kubenswrapper[5018]: E1014 06:51:05.605140 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:05 crc kubenswrapper[5018]: E1014 06:51:05.605207 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:05 crc kubenswrapper[5018]: E1014 06:51:05.605286 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.684560 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.684653 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.684675 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.684698 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.684715 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.788743 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.788789 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.788806 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.788830 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.788847 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.892152 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.892252 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.892272 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.892297 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.892314 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.995433 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.995504 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.995525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.995551 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:05 crc kubenswrapper[5018]: I1014 06:51:05.995570 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:05Z","lastTransitionTime":"2025-10-14T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.098601 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.098724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.098748 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.098779 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.098803 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.201310 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.201369 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.201387 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.201410 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.201427 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.304124 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.304179 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.304201 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.304230 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.304251 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.406893 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.406966 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.406991 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.407020 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.407042 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.510496 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.510590 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.510607 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.510656 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.510678 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.613095 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.613186 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.613205 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.613230 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.613248 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.715821 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.715907 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.715933 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.715962 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.715984 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.818861 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.818915 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.818932 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.818954 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.818972 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.921741 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.921809 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.921828 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.921854 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:06 crc kubenswrapper[5018]: I1014 06:51:06.921872 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:06Z","lastTransitionTime":"2025-10-14T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.024967 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.025024 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.025043 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.025066 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.025085 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.128089 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.128177 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.128194 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.128220 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.128237 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.230895 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.230975 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.230993 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.231021 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.231042 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.340244 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.340301 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.340319 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.340343 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.340361 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.443128 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.443187 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.443205 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.443231 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.443252 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.497204 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.497420 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.497463 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.497512 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.497466438 +0000 UTC m=+148.081513115 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.497586 5018 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.497664 5018 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.497769 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.497739016 +0000 UTC m=+148.081785693 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.497814 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.497794908 +0000 UTC m=+148.081841575 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.546003 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.546075 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.546098 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.546128 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.546151 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.598392 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.598553 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598711 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598757 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598780 5018 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598802 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598841 5018 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598859 5018 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598875 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.598848829 +0000 UTC m=+148.182895496 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.598922 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.5988984 +0000 UTC m=+148.182945067 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.604299 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.604319 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.604381 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.604396 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.604550 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.604786 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.604984 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:07 crc kubenswrapper[5018]: E1014 06:51:07.605206 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.649236 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.649298 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.649318 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.649342 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.649360 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.753036 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.753104 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.753128 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.753158 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.753180 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.855974 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.856050 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.856077 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.856107 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.856131 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.959078 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.959147 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.959164 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.959186 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:07 crc kubenswrapper[5018]: I1014 06:51:07.959200 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:07Z","lastTransitionTime":"2025-10-14T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.061544 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.061603 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.061656 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.061683 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.061702 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.163894 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.163964 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.163983 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.164008 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.164027 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.266443 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.266489 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.266501 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.266517 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.266528 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.369030 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.369070 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.369081 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.369096 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.369107 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.472146 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.472202 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.472220 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.472246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.472263 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.575242 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.575305 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.575322 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.575350 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.575369 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.678740 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.678809 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.678827 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.678851 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.678869 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.781953 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.782025 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.782049 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.782076 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.782098 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.885153 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.885233 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.885257 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.885284 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.885303 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.988404 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.988478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.988500 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.988531 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:08 crc kubenswrapper[5018]: I1014 06:51:08.988554 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:08Z","lastTransitionTime":"2025-10-14T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.091485 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.091553 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.091570 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.091592 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.091609 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.194130 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.194411 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.194423 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.194439 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.194452 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.297044 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.297112 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.297130 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.297155 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.297173 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.400688 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.400738 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.400751 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.400768 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.400778 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.505430 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.505488 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.505500 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.505520 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.505534 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.604548 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.604648 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.604699 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:09 crc kubenswrapper[5018]: E1014 06:51:09.604865 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.604915 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:09 crc kubenswrapper[5018]: E1014 06:51:09.605086 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:09 crc kubenswrapper[5018]: E1014 06:51:09.605141 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:09 crc kubenswrapper[5018]: E1014 06:51:09.605475 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.608816 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.608858 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.608876 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.608900 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.608916 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.711700 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.711728 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.711736 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.711750 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.711761 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.814790 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.814859 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.814883 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.814912 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.814934 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.917666 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.917710 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.917722 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.917740 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:09 crc kubenswrapper[5018]: I1014 06:51:09.917752 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:09Z","lastTransitionTime":"2025-10-14T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.020204 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.020249 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.020260 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.020277 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.020288 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.124221 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.124283 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.124302 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.124326 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.124344 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.227318 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.227375 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.227393 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.227422 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.227443 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.330090 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.330153 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.330171 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.330193 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.330210 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.432886 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.432944 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.432962 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.432989 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.433007 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.535552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.535663 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.535690 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.535724 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.535748 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.638581 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.638709 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.638737 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.638769 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.638790 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.742216 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.742275 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.742291 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.742315 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.742331 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.845555 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.845654 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.845673 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.845701 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.845722 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.948194 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.948246 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.948264 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.948286 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:10 crc kubenswrapper[5018]: I1014 06:51:10.948303 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:10Z","lastTransitionTime":"2025-10-14T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.050699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.050754 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.050766 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.050782 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.050794 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.154021 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.154069 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.154081 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.154099 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.154115 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.256647 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.256699 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.256716 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.256738 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.256755 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.360138 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.360212 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.360232 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.360263 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.360283 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.462350 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.462388 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.462400 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.462416 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.462428 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.570405 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.570503 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.570525 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.570552 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.570576 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.604953 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.605019 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.604969 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:11 crc kubenswrapper[5018]: E1014 06:51:11.605167 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:11 crc kubenswrapper[5018]: E1014 06:51:11.605270 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.605324 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:11 crc kubenswrapper[5018]: E1014 06:51:11.605542 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:11 crc kubenswrapper[5018]: E1014 06:51:11.605660 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.675132 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.675197 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.675214 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.675238 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.675256 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.778650 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.778722 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.778741 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.778765 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.778783 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.881891 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.881945 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.881962 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.881988 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.882005 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.937339 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.937385 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.937402 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.937478 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 06:51:11 crc kubenswrapper[5018]: I1014 06:51:11.937495 5018 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T06:51:11Z","lastTransitionTime":"2025-10-14T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.007138 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj"] Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.007754 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.011155 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.011286 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.011375 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.011749 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.091598 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.09158382 podStartE2EDuration="1m9.09158382s" podCreationTimestamp="2025-10-14 06:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.065969462 +0000 UTC m=+88.650016099" watchObservedRunningTime="2025-10-14 06:51:12.09158382 +0000 UTC m=+88.675630457" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.111141 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-r7ftx" podStartSLOduration=67.111124683 podStartE2EDuration="1m7.111124683s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.110924517 +0000 UTC m=+88.694971184" watchObservedRunningTime="2025-10-14 06:51:12.111124683 +0000 UTC m=+88.695171320" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.154384 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.154352178 podStartE2EDuration="41.154352178s" podCreationTimestamp="2025-10-14 06:50:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.154337537 +0000 UTC m=+88.738384164" watchObservedRunningTime="2025-10-14 06:51:12.154352178 +0000 UTC m=+88.738398805" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.156585 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/384b6a1f-de9f-4fd0-885d-7e3c494e2001-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.156670 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.156714 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.156737 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/384b6a1f-de9f-4fd0-885d-7e3c494e2001-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.156816 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384b6a1f-de9f-4fd0-885d-7e3c494e2001-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.182027 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-db8xr" podStartSLOduration=68.182007175 podStartE2EDuration="1m8.182007175s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.1818449 +0000 UTC m=+88.765891527" watchObservedRunningTime="2025-10-14 06:51:12.182007175 +0000 UTC m=+88.766053802" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.193579 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podStartSLOduration=68.193560237 podStartE2EDuration="1m8.193560237s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.193139065 +0000 UTC m=+88.777185682" watchObservedRunningTime="2025-10-14 06:51:12.193560237 +0000 UTC m=+88.777606864" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.222520 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fzdc9" podStartSLOduration=67.222494491 podStartE2EDuration="1m7.222494491s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.210263088 +0000 UTC m=+88.794309715" watchObservedRunningTime="2025-10-14 06:51:12.222494491 +0000 UTC m=+88.806541118" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.249163 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.249121238 podStartE2EDuration="1m8.249121238s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.248444078 +0000 UTC m=+88.832490705" watchObservedRunningTime="2025-10-14 06:51:12.249121238 +0000 UTC m=+88.833167905" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.249466 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxzt" podStartSLOduration=67.249454008 podStartE2EDuration="1m7.249454008s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.22314138 +0000 UTC m=+88.807188007" watchObservedRunningTime="2025-10-14 06:51:12.249454008 +0000 UTC m=+88.833500685" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257254 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257293 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/384b6a1f-de9f-4fd0-885d-7e3c494e2001-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257341 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384b6a1f-de9f-4fd0-885d-7e3c494e2001-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257363 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/384b6a1f-de9f-4fd0-885d-7e3c494e2001-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257383 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257401 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.257437 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/384b6a1f-de9f-4fd0-885d-7e3c494e2001-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.258225 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/384b6a1f-de9f-4fd0-885d-7e3c494e2001-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.268080 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/384b6a1f-de9f-4fd0-885d-7e3c494e2001-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.277104 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/384b6a1f-de9f-4fd0-885d-7e3c494e2001-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s97mj\" (UID: \"384b6a1f-de9f-4fd0-885d-7e3c494e2001\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: I1014 06:51:12.356099 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" Oct 14 06:51:12 crc kubenswrapper[5018]: W1014 06:51:12.367608 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod384b6a1f_de9f_4fd0_885d_7e3c494e2001.slice/crio-30415c06357d4d3129eb998f61fd0855a893dba1b36df9057514ef740bea81c9 WatchSource:0}: Error finding container 30415c06357d4d3129eb998f61fd0855a893dba1b36df9057514ef740bea81c9: Status 404 returned error can't find the container with id 30415c06357d4d3129eb998f61fd0855a893dba1b36df9057514ef740bea81c9 Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.182610 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" event={"ID":"384b6a1f-de9f-4fd0-885d-7e3c494e2001","Type":"ContainerStarted","Data":"7a8c52c6b39c3edfeb933c487240cbebbaa0e9785905fb0b94a865f3ca6a8e03"} Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.182724 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" event={"ID":"384b6a1f-de9f-4fd0-885d-7e3c494e2001","Type":"ContainerStarted","Data":"30415c06357d4d3129eb998f61fd0855a893dba1b36df9057514ef740bea81c9"} Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.199083 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pwgtv" podStartSLOduration=69.199056422 podStartE2EDuration="1m9.199056422s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:12.343781445 +0000 UTC m=+88.927828072" watchObservedRunningTime="2025-10-14 06:51:13.199056422 +0000 UTC m=+89.783103069" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.604598 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.604796 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.605317 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:13 crc kubenswrapper[5018]: E1014 06:51:13.605408 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:13 crc kubenswrapper[5018]: E1014 06:51:13.605740 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:13 crc kubenswrapper[5018]: E1014 06:51:13.606013 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.606125 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:13 crc kubenswrapper[5018]: E1014 06:51:13.606261 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.621935 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s97mj" podStartSLOduration=68.621906593 podStartE2EDuration="1m8.621906593s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:13.200131733 +0000 UTC m=+89.784178400" watchObservedRunningTime="2025-10-14 06:51:13.621906593 +0000 UTC m=+90.205953230" Oct 14 06:51:13 crc kubenswrapper[5018]: I1014 06:51:13.622318 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 14 06:51:14 crc kubenswrapper[5018]: I1014 06:51:14.621613 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 14 06:51:14 crc kubenswrapper[5018]: I1014 06:51:14.641519 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.6414934140000002 podStartE2EDuration="1.641493414s" podCreationTimestamp="2025-10-14 06:51:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:14.641293618 +0000 UTC m=+91.225340275" watchObservedRunningTime="2025-10-14 06:51:14.641493414 +0000 UTC m=+91.225540081" Oct 14 06:51:15 crc kubenswrapper[5018]: I1014 06:51:15.604954 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:15 crc kubenswrapper[5018]: I1014 06:51:15.605014 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:15 crc kubenswrapper[5018]: I1014 06:51:15.604951 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:15 crc kubenswrapper[5018]: E1014 06:51:15.605066 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:15 crc kubenswrapper[5018]: I1014 06:51:15.605023 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:15 crc kubenswrapper[5018]: E1014 06:51:15.605181 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:15 crc kubenswrapper[5018]: E1014 06:51:15.605389 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:15 crc kubenswrapper[5018]: E1014 06:51:15.605544 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:17 crc kubenswrapper[5018]: I1014 06:51:17.605087 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:17 crc kubenswrapper[5018]: I1014 06:51:17.605157 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:17 crc kubenswrapper[5018]: I1014 06:51:17.605166 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:17 crc kubenswrapper[5018]: I1014 06:51:17.605118 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:17 crc kubenswrapper[5018]: E1014 06:51:17.605376 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:17 crc kubenswrapper[5018]: E1014 06:51:17.605498 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:17 crc kubenswrapper[5018]: E1014 06:51:17.605669 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:17 crc kubenswrapper[5018]: E1014 06:51:17.605867 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:19 crc kubenswrapper[5018]: I1014 06:51:19.604543 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:19 crc kubenswrapper[5018]: I1014 06:51:19.604611 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:19 crc kubenswrapper[5018]: I1014 06:51:19.605257 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:19 crc kubenswrapper[5018]: I1014 06:51:19.605339 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:19 crc kubenswrapper[5018]: E1014 06:51:19.605778 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:19 crc kubenswrapper[5018]: E1014 06:51:19.605964 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:19 crc kubenswrapper[5018]: I1014 06:51:19.606126 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 06:51:19 crc kubenswrapper[5018]: E1014 06:51:19.606436 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:51:19 crc kubenswrapper[5018]: E1014 06:51:19.606594 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:19 crc kubenswrapper[5018]: E1014 06:51:19.606821 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:21 crc kubenswrapper[5018]: I1014 06:51:21.603884 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:21 crc kubenswrapper[5018]: E1014 06:51:21.603998 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:21 crc kubenswrapper[5018]: I1014 06:51:21.603885 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:21 crc kubenswrapper[5018]: I1014 06:51:21.603884 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:21 crc kubenswrapper[5018]: E1014 06:51:21.604082 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:21 crc kubenswrapper[5018]: I1014 06:51:21.604055 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:21 crc kubenswrapper[5018]: E1014 06:51:21.604251 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:21 crc kubenswrapper[5018]: E1014 06:51:21.604373 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:23 crc kubenswrapper[5018]: I1014 06:51:23.595601 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.595895 5018 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.595968 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs podName:2530b46d-5f3a-4495-af23-3ad85327c99c nodeName:}" failed. No retries permitted until 2025-10-14 06:52:27.595944904 +0000 UTC m=+164.179991561 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs") pod "network-metrics-daemon-gl9fg" (UID: "2530b46d-5f3a-4495-af23-3ad85327c99c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 06:51:23 crc kubenswrapper[5018]: I1014 06:51:23.604459 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:23 crc kubenswrapper[5018]: I1014 06:51:23.604500 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:23 crc kubenswrapper[5018]: I1014 06:51:23.604514 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.604683 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:23 crc kubenswrapper[5018]: I1014 06:51:23.604737 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.604952 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.605334 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:23 crc kubenswrapper[5018]: E1014 06:51:23.605459 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:24 crc kubenswrapper[5018]: I1014 06:51:24.626548 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.626523161 podStartE2EDuration="10.626523161s" podCreationTimestamp="2025-10-14 06:51:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:24.625069459 +0000 UTC m=+101.209116126" watchObservedRunningTime="2025-10-14 06:51:24.626523161 +0000 UTC m=+101.210569818" Oct 14 06:51:25 crc kubenswrapper[5018]: I1014 06:51:25.604859 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:25 crc kubenswrapper[5018]: I1014 06:51:25.604912 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:25 crc kubenswrapper[5018]: I1014 06:51:25.604954 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:25 crc kubenswrapper[5018]: E1014 06:51:25.605690 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:25 crc kubenswrapper[5018]: E1014 06:51:25.605416 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:25 crc kubenswrapper[5018]: I1014 06:51:25.604994 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:25 crc kubenswrapper[5018]: E1014 06:51:25.605758 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:25 crc kubenswrapper[5018]: E1014 06:51:25.605878 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:27 crc kubenswrapper[5018]: I1014 06:51:27.604281 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:27 crc kubenswrapper[5018]: I1014 06:51:27.604356 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:27 crc kubenswrapper[5018]: I1014 06:51:27.604562 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:27 crc kubenswrapper[5018]: E1014 06:51:27.604813 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:27 crc kubenswrapper[5018]: I1014 06:51:27.604883 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:27 crc kubenswrapper[5018]: E1014 06:51:27.605093 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:27 crc kubenswrapper[5018]: E1014 06:51:27.605127 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:27 crc kubenswrapper[5018]: E1014 06:51:27.605201 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:29 crc kubenswrapper[5018]: I1014 06:51:29.604730 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:29 crc kubenswrapper[5018]: I1014 06:51:29.604817 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:29 crc kubenswrapper[5018]: I1014 06:51:29.604845 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:29 crc kubenswrapper[5018]: I1014 06:51:29.604916 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:29 crc kubenswrapper[5018]: E1014 06:51:29.605051 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:29 crc kubenswrapper[5018]: E1014 06:51:29.605223 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:29 crc kubenswrapper[5018]: E1014 06:51:29.605438 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:29 crc kubenswrapper[5018]: E1014 06:51:29.605566 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:31 crc kubenswrapper[5018]: I1014 06:51:31.604677 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:31 crc kubenswrapper[5018]: I1014 06:51:31.604783 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:31 crc kubenswrapper[5018]: E1014 06:51:31.604927 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:31 crc kubenswrapper[5018]: I1014 06:51:31.604734 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:31 crc kubenswrapper[5018]: I1014 06:51:31.605039 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:31 crc kubenswrapper[5018]: E1014 06:51:31.605075 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:31 crc kubenswrapper[5018]: E1014 06:51:31.605199 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:31 crc kubenswrapper[5018]: E1014 06:51:31.605443 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:32 crc kubenswrapper[5018]: I1014 06:51:32.605245 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 06:51:32 crc kubenswrapper[5018]: E1014 06:51:32.605565 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c2zpf_openshift-ovn-kubernetes(99edde9a-4116-4bdf-84ff-863e9e6ed55e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" Oct 14 06:51:33 crc kubenswrapper[5018]: I1014 06:51:33.604521 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:33 crc kubenswrapper[5018]: I1014 06:51:33.604528 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:33 crc kubenswrapper[5018]: I1014 06:51:33.604862 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:33 crc kubenswrapper[5018]: E1014 06:51:33.604742 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:33 crc kubenswrapper[5018]: I1014 06:51:33.604543 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:33 crc kubenswrapper[5018]: E1014 06:51:33.605087 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:33 crc kubenswrapper[5018]: E1014 06:51:33.605202 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:33 crc kubenswrapper[5018]: E1014 06:51:33.605299 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:35 crc kubenswrapper[5018]: I1014 06:51:35.604760 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:35 crc kubenswrapper[5018]: E1014 06:51:35.605473 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:35 crc kubenswrapper[5018]: I1014 06:51:35.604936 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:35 crc kubenswrapper[5018]: E1014 06:51:35.605704 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:35 crc kubenswrapper[5018]: I1014 06:51:35.604906 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:35 crc kubenswrapper[5018]: I1014 06:51:35.604963 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:35 crc kubenswrapper[5018]: E1014 06:51:35.605920 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:35 crc kubenswrapper[5018]: E1014 06:51:35.605982 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:37 crc kubenswrapper[5018]: I1014 06:51:37.604429 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:37 crc kubenswrapper[5018]: I1014 06:51:37.604547 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:37 crc kubenswrapper[5018]: I1014 06:51:37.604595 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:37 crc kubenswrapper[5018]: E1014 06:51:37.604755 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:37 crc kubenswrapper[5018]: I1014 06:51:37.605053 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:37 crc kubenswrapper[5018]: E1014 06:51:37.605294 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:37 crc kubenswrapper[5018]: E1014 06:51:37.605376 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:37 crc kubenswrapper[5018]: E1014 06:51:37.605535 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.283445 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/1.log" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.284230 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/0.log" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.284304 5018 generic.go:334] "Generic (PLEG): container finished" podID="f1aa622b-96ac-42aa-8bd1-8fa174e65602" containerID="7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a" exitCode=1 Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.284352 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerDied","Data":"7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a"} Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.284408 5018 scope.go:117] "RemoveContainer" containerID="e4655fd4206a556c2581fda39662e35a382a8aff850ebcd6b02fdae6808bc8fb" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.285688 5018 scope.go:117] "RemoveContainer" containerID="7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a" Oct 14 06:51:39 crc kubenswrapper[5018]: E1014 06:51:39.286086 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-r7ftx_openshift-multus(f1aa622b-96ac-42aa-8bd1-8fa174e65602)\"" pod="openshift-multus/multus-r7ftx" podUID="f1aa622b-96ac-42aa-8bd1-8fa174e65602" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.604386 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.604435 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.604459 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:39 crc kubenswrapper[5018]: I1014 06:51:39.604407 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:39 crc kubenswrapper[5018]: E1014 06:51:39.604570 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:39 crc kubenswrapper[5018]: E1014 06:51:39.604720 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:39 crc kubenswrapper[5018]: E1014 06:51:39.604821 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:39 crc kubenswrapper[5018]: E1014 06:51:39.604995 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:40 crc kubenswrapper[5018]: I1014 06:51:40.289444 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/1.log" Oct 14 06:51:41 crc kubenswrapper[5018]: I1014 06:51:41.604797 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:41 crc kubenswrapper[5018]: I1014 06:51:41.604826 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:41 crc kubenswrapper[5018]: I1014 06:51:41.604872 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:41 crc kubenswrapper[5018]: I1014 06:51:41.604964 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:41 crc kubenswrapper[5018]: E1014 06:51:41.605136 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:41 crc kubenswrapper[5018]: E1014 06:51:41.605245 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:41 crc kubenswrapper[5018]: E1014 06:51:41.605376 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:41 crc kubenswrapper[5018]: E1014 06:51:41.605532 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:43 crc kubenswrapper[5018]: I1014 06:51:43.604297 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:43 crc kubenswrapper[5018]: I1014 06:51:43.604352 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:43 crc kubenswrapper[5018]: I1014 06:51:43.604286 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:43 crc kubenswrapper[5018]: I1014 06:51:43.604441 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:43 crc kubenswrapper[5018]: E1014 06:51:43.604534 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:43 crc kubenswrapper[5018]: E1014 06:51:43.604607 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:43 crc kubenswrapper[5018]: E1014 06:51:43.604764 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:43 crc kubenswrapper[5018]: E1014 06:51:43.604935 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:44 crc kubenswrapper[5018]: E1014 06:51:44.597194 5018 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 14 06:51:44 crc kubenswrapper[5018]: E1014 06:51:44.733311 5018 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 06:51:45 crc kubenswrapper[5018]: I1014 06:51:45.604147 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:45 crc kubenswrapper[5018]: I1014 06:51:45.604224 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:45 crc kubenswrapper[5018]: I1014 06:51:45.604223 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:45 crc kubenswrapper[5018]: I1014 06:51:45.604229 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:45 crc kubenswrapper[5018]: E1014 06:51:45.604287 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:45 crc kubenswrapper[5018]: E1014 06:51:45.604325 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:45 crc kubenswrapper[5018]: E1014 06:51:45.604445 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:45 crc kubenswrapper[5018]: E1014 06:51:45.604687 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:46 crc kubenswrapper[5018]: I1014 06:51:46.605498 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.315878 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/3.log" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.318599 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerStarted","Data":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.319041 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.361439 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podStartSLOduration=102.361409348 podStartE2EDuration="1m42.361409348s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:51:47.359549354 +0000 UTC m=+123.943596001" watchObservedRunningTime="2025-10-14 06:51:47.361409348 +0000 UTC m=+123.945456025" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.518957 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gl9fg"] Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.519068 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:47 crc kubenswrapper[5018]: E1014 06:51:47.519170 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.604196 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.604235 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:47 crc kubenswrapper[5018]: E1014 06:51:47.604359 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:47 crc kubenswrapper[5018]: I1014 06:51:47.604394 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:47 crc kubenswrapper[5018]: E1014 06:51:47.604508 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:47 crc kubenswrapper[5018]: E1014 06:51:47.604577 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:49 crc kubenswrapper[5018]: I1014 06:51:49.604858 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:49 crc kubenswrapper[5018]: I1014 06:51:49.604899 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:49 crc kubenswrapper[5018]: E1014 06:51:49.605409 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:49 crc kubenswrapper[5018]: I1014 06:51:49.604998 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:49 crc kubenswrapper[5018]: I1014 06:51:49.605425 5018 scope.go:117] "RemoveContainer" containerID="7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a" Oct 14 06:51:49 crc kubenswrapper[5018]: E1014 06:51:49.605517 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:49 crc kubenswrapper[5018]: I1014 06:51:49.604936 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:49 crc kubenswrapper[5018]: E1014 06:51:49.605723 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:49 crc kubenswrapper[5018]: E1014 06:51:49.605940 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:49 crc kubenswrapper[5018]: E1014 06:51:49.734441 5018 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 06:51:50 crc kubenswrapper[5018]: I1014 06:51:50.333117 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/1.log" Oct 14 06:51:50 crc kubenswrapper[5018]: I1014 06:51:50.333197 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerStarted","Data":"0eb191c53d3c98d15cfe80fae0123a7b7d712dc828106085b6059d32258a3d98"} Oct 14 06:51:51 crc kubenswrapper[5018]: I1014 06:51:51.604973 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:51 crc kubenswrapper[5018]: I1014 06:51:51.605019 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:51 crc kubenswrapper[5018]: I1014 06:51:51.605086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:51 crc kubenswrapper[5018]: I1014 06:51:51.605086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:51 crc kubenswrapper[5018]: E1014 06:51:51.605164 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:51 crc kubenswrapper[5018]: E1014 06:51:51.605338 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:51 crc kubenswrapper[5018]: E1014 06:51:51.605455 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:51 crc kubenswrapper[5018]: E1014 06:51:51.605564 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:53 crc kubenswrapper[5018]: I1014 06:51:53.604266 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:53 crc kubenswrapper[5018]: E1014 06:51:53.604817 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 06:51:53 crc kubenswrapper[5018]: I1014 06:51:53.604412 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:53 crc kubenswrapper[5018]: E1014 06:51:53.604934 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 06:51:53 crc kubenswrapper[5018]: I1014 06:51:53.604459 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:53 crc kubenswrapper[5018]: E1014 06:51:53.605025 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 06:51:53 crc kubenswrapper[5018]: I1014 06:51:53.604406 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:53 crc kubenswrapper[5018]: E1014 06:51:53.605118 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gl9fg" podUID="2530b46d-5f3a-4495-af23-3ad85327c99c" Oct 14 06:51:53 crc kubenswrapper[5018]: I1014 06:51:53.785566 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.604348 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.604407 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.604496 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.604756 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.607142 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.607424 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.607147 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.607882 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.607994 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 14 06:51:55 crc kubenswrapper[5018]: I1014 06:51:55.608063 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.860488 5018 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.921009 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dp82w"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.921717 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.925486 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7dthh"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.927221 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.927423 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.927709 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.928913 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.929447 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.930519 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f9c54"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.931336 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.941544 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.943815 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.944114 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.944292 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951170 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951204 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951407 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951520 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951560 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951412 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951745 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951921 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.952093 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.952225 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.951522 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.952377 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.952551 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.952670 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953415 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953597 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953696 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953798 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953940 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.953980 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.954130 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.954322 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.954382 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.954575 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.954817 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.956492 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.956927 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957096 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957165 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957257 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957288 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957468 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957685 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.957692 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.958429 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.960406 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.962723 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.962901 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.965651 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.967002 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.967601 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.967915 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.968065 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.968425 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.968705 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.968860 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.970014 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.970176 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.970322 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.970870 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.971483 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.972833 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.974799 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vc4k4"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.975783 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzjvj"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.976537 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.977228 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.977916 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.978026 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.978229 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.978275 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.978937 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.978944 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.979604 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.979736 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.979911 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.980323 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.980539 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5hjlc"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.980820 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.980861 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.980834 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.981832 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.982516 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.982543 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.982849 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.985530 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.986049 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rgqhn"] Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.986427 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.986750 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.990257 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.990527 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.990690 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.990765 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.990917 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.991194 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.991274 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.991341 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 06:52:02 crc kubenswrapper[5018]: I1014 06:52:02.991406 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.013413 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-zvw8c"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.014921 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015605 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015691 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015762 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015872 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015898 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015931 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.015944 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.016555 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.023016 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.023895 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.024314 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.024503 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.024588 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.024794 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025118 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025240 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025469 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025598 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025823 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.025930 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.026106 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-p9sgv"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.026331 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.026704 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027061 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027192 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027324 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027390 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027527 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.027715 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.028000 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.028182 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.028481 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.033029 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.033084 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.033249 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.033398 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dp82w"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.039808 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.040302 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.040557 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.042519 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.043141 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.043879 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.045711 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.047401 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.050052 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.052431 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.052779 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.053509 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.053518 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.053603 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.053865 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058667 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8241e590-8634-48c7-b07c-3ef44b4d2f85-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058699 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-dir\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058720 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czt2r\" (UniqueName: \"kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058739 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058756 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058771 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-policies\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058787 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058813 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-serving-cert\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058829 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-config\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058854 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8241e590-8634-48c7-b07c-3ef44b4d2f85-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058874 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058903 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7cd77e7d-730d-4259-b784-4a7b8add26e0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058919 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058934 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-encryption-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058950 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058968 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjg4c\" (UniqueName: \"kubernetes.io/projected/c6385ef6-72b4-49f4-b61b-546be9da202c-kube-api-access-wjg4c\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.058996 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059012 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059040 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059053 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-encryption-config\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059066 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-client\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059080 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-image-import-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059098 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059113 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059129 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059144 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059160 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-client\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059174 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-images\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059190 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059204 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-node-pullsecrets\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059221 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit-dir\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059236 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-serving-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059250 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059264 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bcd\" (UniqueName: \"kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059286 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9mj9\" (UniqueName: \"kubernetes.io/projected/7cd77e7d-730d-4259-b784-4a7b8add26e0-kube-api-access-k9mj9\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059300 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q7lm\" (UniqueName: \"kubernetes.io/projected/8241e590-8634-48c7-b07c-3ef44b4d2f85-kube-api-access-7q7lm\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059315 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd47w\" (UniqueName: \"kubernetes.io/projected/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-kube-api-access-kd47w\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059332 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/313736e9-309a-4784-a1b1-c7412d322eb5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059349 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059364 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv9f9\" (UniqueName: \"kubernetes.io/projected/313736e9-309a-4784-a1b1-c7412d322eb5-kube-api-access-dv9f9\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059451 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-serving-cert\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059506 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313736e9-309a-4784-a1b1-c7412d322eb5-serving-cert\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059525 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.059542 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.061805 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.062342 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.063773 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.063923 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.066156 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.066611 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.077827 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.078929 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.079305 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.079491 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.081814 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r744g"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.085076 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.085485 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.086405 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.086584 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.087338 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.088160 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.092197 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.092313 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8mzkh"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.092829 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.092919 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.093017 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.095394 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vc4k4"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.097099 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7dthh"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.097616 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.099039 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzjvj"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.102496 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.102757 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-p9sgv"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.103733 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.105750 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.105931 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.108588 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.111275 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.112651 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.114488 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.115212 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.115933 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dh52n"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.116421 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.117002 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.117407 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.118416 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.118807 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.119515 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hl8xt"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.119864 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.121166 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-mhqlx"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.121619 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.122558 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.123888 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.124844 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.126483 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4j2t6"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.127417 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.128142 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x2z8l"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.128909 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.129159 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.130551 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rgqhn"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.131847 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.132677 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.133633 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.134643 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.135558 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.136974 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.137567 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.138813 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zvw8c"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.140141 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.141417 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.142900 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f9c54"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.143505 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.143917 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.145164 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.145993 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r744g"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.147056 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.148262 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8mzkh"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.149383 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dh52n"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.150957 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.152066 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x2z8l"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.154599 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.156019 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hl8xt"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.157123 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4j2t6"] Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160561 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-serving-cert\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160593 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9c9584-72fd-4b26-98f9-917bb1978c7b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160612 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-images\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160641 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f29c068e-309c-478a-a64d-8698c73028f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160663 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29c068e-309c-478a-a64d-8698c73028f2-config\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-machine-approver-tls\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160693 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313736e9-309a-4784-a1b1-c7412d322eb5-serving-cert\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160708 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160723 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160738 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0f1ed6a-5861-471c-979b-76cfae275d2f-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160760 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8241e590-8634-48c7-b07c-3ef44b4d2f85-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160777 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-dir\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160794 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czt2r\" (UniqueName: \"kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160809 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160824 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160845 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2245a631-4bde-4030-8318-2c7b57171224-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160862 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222c5fe-b248-425d-a810-83e03907f6a7-serving-cert\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160885 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-policies\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160900 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160916 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8djll\" (UniqueName: \"kubernetes.io/projected/54e62cc6-54c6-4df2-b349-39bf04d702bc-kube-api-access-8djll\") pod \"downloads-7954f5f757-zvw8c\" (UID: \"54e62cc6-54c6-4df2-b349-39bf04d702bc\") " pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160930 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-545rn\" (UniqueName: \"kubernetes.io/projected/6559b957-5c56-410a-86d6-b75422bf82b3-kube-api-access-545rn\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160947 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-serving-cert\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160962 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160977 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c12913c-de90-4dd3-a247-8a93e53565dc-config\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.160994 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-metrics-certs\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161019 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-config\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161038 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49c16b2d-7e27-4017-a41e-ab204084c2a7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161056 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8241e590-8634-48c7-b07c-3ef44b4d2f85-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161074 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvlq9\" (UniqueName: \"kubernetes.io/projected/68755a31-39a6-40e9-9317-1876c7393527-kube-api-access-zvlq9\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161092 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zxnx\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-kube-api-access-4zxnx\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161109 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljb8c\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-kube-api-access-ljb8c\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161249 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161267 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161282 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c12913c-de90-4dd3-a247-8a93e53565dc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161351 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68755a31-39a6-40e9-9317-1876c7393527-proxy-tls\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161384 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-config\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161399 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrpkr\" (UniqueName: \"kubernetes.io/projected/ea6489fd-1b53-4ef0-9561-7691dd063b0c-kube-api-access-rrpkr\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161415 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f9c9584-72fd-4b26-98f9-917bb1978c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161428 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-auth-proxy-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161445 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7cd77e7d-730d-4259-b784-4a7b8add26e0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161461 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161476 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-encryption-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161492 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161508 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjg4c\" (UniqueName: \"kubernetes.io/projected/c6385ef6-72b4-49f4-b61b-546be9da202c-kube-api-access-wjg4c\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161522 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161537 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-service-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gd5x\" (UniqueName: \"kubernetes.io/projected/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-kube-api-access-6gd5x\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161567 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161585 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.161745 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-dir\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.162565 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.162767 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-config\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.162849 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163013 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163017 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163077 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163103 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163132 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-srv-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163149 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163167 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-default-certificate\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163193 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-client\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163210 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-image-import-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163227 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c01fd1f8-596f-445e-a209-c835c6d754f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163251 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163277 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-encryption-config\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163296 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163315 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163334 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163350 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c12913c-de90-4dd3-a247-8a93e53565dc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163917 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5mdx\" (UniqueName: \"kubernetes.io/projected/49c16b2d-7e27-4017-a41e-ab204084c2a7-kube-api-access-p5mdx\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.163987 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164041 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-client\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164125 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qzsb\" (UniqueName: \"kubernetes.io/projected/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-kube-api-access-4qzsb\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164162 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-images\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164193 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164221 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f29c068e-309c-478a-a64d-8698c73028f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164251 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-node-pullsecrets\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164277 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0f1ed6a-5861-471c-979b-76cfae275d2f-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164418 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit-dir\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164464 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-serving-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164514 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bcd\" (UniqueName: \"kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164549 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9c9584-72fd-4b26-98f9-917bb1978c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164566 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-config\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164586 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-trusted-ca\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164643 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd47w\" (UniqueName: \"kubernetes.io/projected/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-kube-api-access-kd47w\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164669 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116c38b3-dbff-4f91-b057-93cc6ae4bb96-serving-cert\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164689 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc8mn\" (UniqueName: \"kubernetes.io/projected/116c38b3-dbff-4f91-b057-93cc6ae4bb96-kube-api-access-kc8mn\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164708 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-stats-auth\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164728 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9mj9\" (UniqueName: \"kubernetes.io/projected/7cd77e7d-730d-4259-b784-4a7b8add26e0-kube-api-access-k9mj9\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164768 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q7lm\" (UniqueName: \"kubernetes.io/projected/8241e590-8634-48c7-b07c-3ef44b4d2f85-kube-api-access-7q7lm\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164790 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/313736e9-309a-4784-a1b1-c7412d322eb5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164809 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2245a631-4bde-4030-8318-2c7b57171224-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164839 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6559b957-5c56-410a-86d6-b75422bf82b3-service-ca-bundle\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164958 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.164987 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv9f9\" (UniqueName: \"kubernetes.io/projected/313736e9-309a-4784-a1b1-c7412d322eb5-kube-api-access-dv9f9\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.165027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-metrics-tls\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.165047 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c01fd1f8-596f-445e-a209-c835c6d754f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.165066 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dt4l\" (UniqueName: \"kubernetes.io/projected/c01fd1f8-596f-445e-a209-c835c6d754f0-kube-api-access-8dt4l\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.165082 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j26c\" (UniqueName: \"kubernetes.io/projected/d222c5fe-b248-425d-a810-83e03907f6a7-kube-api-access-8j26c\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.165850 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.167434 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.167998 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-serving-cert\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.168175 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8241e590-8634-48c7-b07c-3ef44b4d2f85-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.168271 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-image-import-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.168692 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7cd77e7d-730d-4259-b784-4a7b8add26e0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.169294 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.169474 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7cd77e7d-730d-4259-b784-4a7b8add26e0-images\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.169794 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.170041 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.170206 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-node-pullsecrets\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.170305 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-audit-dir\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.170429 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.170853 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-serving-ca\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.171665 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.172244 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-serving-cert\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.172491 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-etcd-client\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.172852 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-encryption-config\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.172951 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-encryption-config\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.173345 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.173485 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.173558 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.173777 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/313736e9-309a-4784-a1b1-c7412d322eb5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.174084 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.173334 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.174493 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6385ef6-72b4-49f4-b61b-546be9da202c-audit-policies\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.174605 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.175197 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.175234 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8241e590-8634-48c7-b07c-3ef44b4d2f85-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.176532 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.176883 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.177720 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.177832 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.177943 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.178391 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.179740 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c6385ef6-72b4-49f4-b61b-546be9da202c-etcd-client\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.181663 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313736e9-309a-4784-a1b1-c7412d322eb5-serving-cert\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.183207 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.204551 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.223851 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.244666 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.264083 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265614 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5mdx\" (UniqueName: \"kubernetes.io/projected/49c16b2d-7e27-4017-a41e-ab204084c2a7-kube-api-access-p5mdx\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265759 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c12913c-de90-4dd3-a247-8a93e53565dc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265797 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265821 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qzsb\" (UniqueName: \"kubernetes.io/projected/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-kube-api-access-4qzsb\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265848 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f29c068e-309c-478a-a64d-8698c73028f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265872 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0f1ed6a-5861-471c-979b-76cfae275d2f-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265902 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9c9584-72fd-4b26-98f9-917bb1978c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265925 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-config\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.265951 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-trusted-ca\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266118 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc8mn\" (UniqueName: \"kubernetes.io/projected/116c38b3-dbff-4f91-b057-93cc6ae4bb96-kube-api-access-kc8mn\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266161 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-stats-auth\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266210 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116c38b3-dbff-4f91-b057-93cc6ae4bb96-serving-cert\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266238 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6559b957-5c56-410a-86d6-b75422bf82b3-service-ca-bundle\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266265 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2245a631-4bde-4030-8318-2c7b57171224-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266291 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c01fd1f8-596f-445e-a209-c835c6d754f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266315 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dt4l\" (UniqueName: \"kubernetes.io/projected/c01fd1f8-596f-445e-a209-c835c6d754f0-kube-api-access-8dt4l\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266340 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j26c\" (UniqueName: \"kubernetes.io/projected/d222c5fe-b248-425d-a810-83e03907f6a7-kube-api-access-8j26c\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266374 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-metrics-tls\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266397 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-images\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266418 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f29c068e-309c-478a-a64d-8698c73028f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266439 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29c068e-309c-478a-a64d-8698c73028f2-config\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266460 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-machine-approver-tls\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266489 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9c9584-72fd-4b26-98f9-917bb1978c7b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266512 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0f1ed6a-5861-471c-979b-76cfae275d2f-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266562 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2245a631-4bde-4030-8318-2c7b57171224-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266587 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222c5fe-b248-425d-a810-83e03907f6a7-serving-cert\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266617 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8djll\" (UniqueName: \"kubernetes.io/projected/54e62cc6-54c6-4df2-b349-39bf04d702bc-kube-api-access-8djll\") pod \"downloads-7954f5f757-zvw8c\" (UID: \"54e62cc6-54c6-4df2-b349-39bf04d702bc\") " pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266656 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-545rn\" (UniqueName: \"kubernetes.io/projected/6559b957-5c56-410a-86d6-b75422bf82b3-kube-api-access-545rn\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266684 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266739 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c12913c-de90-4dd3-a247-8a93e53565dc-config\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266760 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-metrics-certs\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266799 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49c16b2d-7e27-4017-a41e-ab204084c2a7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266821 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvlq9\" (UniqueName: \"kubernetes.io/projected/68755a31-39a6-40e9-9317-1876c7393527-kube-api-access-zvlq9\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266844 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zxnx\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-kube-api-access-4zxnx\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266866 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljb8c\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-kube-api-access-ljb8c\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266900 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c12913c-de90-4dd3-a247-8a93e53565dc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266923 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68755a31-39a6-40e9-9317-1876c7393527-proxy-tls\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266955 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrpkr\" (UniqueName: \"kubernetes.io/projected/ea6489fd-1b53-4ef0-9561-7691dd063b0c-kube-api-access-rrpkr\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.266982 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-config\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267003 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-auth-proxy-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267026 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f9c9584-72fd-4b26-98f9-917bb1978c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267055 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-service-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267076 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gd5x\" (UniqueName: \"kubernetes.io/projected/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-kube-api-access-6gd5x\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267098 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267130 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267155 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-srv-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267176 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267198 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-default-certificate\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267218 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c01fd1f8-596f-445e-a209-c835c6d754f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267281 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-trusted-ca\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.267798 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2245a631-4bde-4030-8318-2c7b57171224-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.268471 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-auth-proxy-config\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.268652 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f9c9584-72fd-4b26-98f9-917bb1978c7b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.268704 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6559b957-5c56-410a-86d6-b75422bf82b3-service-ca-bundle\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.268765 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e0f1ed6a-5861-471c-979b-76cfae275d2f-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.268836 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-metrics-tls\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.269388 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d222c5fe-b248-425d-a810-83e03907f6a7-config\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.269550 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-auth-proxy-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.269896 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-config\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.270428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0f1ed6a-5861-471c-979b-76cfae275d2f-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.271052 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-stats-auth\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.271643 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-default-certificate\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.271922 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6559b957-5c56-410a-86d6-b75422bf82b3-metrics-certs\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.271987 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49c16b2d-7e27-4017-a41e-ab204084c2a7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.272543 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-machine-approver-tls\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.272888 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f9c9584-72fd-4b26-98f9-917bb1978c7b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.273129 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d222c5fe-b248-425d-a810-83e03907f6a7-serving-cert\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.273654 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2245a631-4bde-4030-8318-2c7b57171224-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.284352 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.303754 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.324098 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.344096 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.363717 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.383520 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.403476 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.410650 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f29c068e-309c-478a-a64d-8698c73028f2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.424533 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.427822 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29c068e-309c-478a-a64d-8698c73028f2-config\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.443898 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.449410 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c01fd1f8-596f-445e-a209-c835c6d754f0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.463768 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.484629 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.491584 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c01fd1f8-596f-445e-a209-c835c6d754f0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.503886 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.525227 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.544088 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.563912 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.604191 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.624978 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.653116 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.659388 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.664808 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.667545 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-config\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.683851 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.704409 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.712128 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116c38b3-dbff-4f91-b057-93cc6ae4bb96-serving-cert\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.724051 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.730472 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/116c38b3-dbff-4f91-b057-93cc6ae4bb96-service-ca-bundle\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.745478 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.765206 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.785320 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.806389 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.809005 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/68755a31-39a6-40e9-9317-1876c7393527-images\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.824385 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.845291 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.864999 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.869814 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c12913c-de90-4dd3-a247-8a93e53565dc-config\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.884332 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.893910 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/68755a31-39a6-40e9-9317-1876c7393527-proxy-tls\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.904829 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.913447 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c12913c-de90-4dd3-a247-8a93e53565dc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.924924 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.945881 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.964592 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.971607 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-srv-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.984921 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 14 06:52:03 crc kubenswrapper[5018]: I1014 06:52:03.995212 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ea6489fd-1b53-4ef0-9561-7691dd063b0c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.004534 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.024453 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.064720 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.083000 5018 request.go:700] Waited for 1.003367362s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.084813 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.103530 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.124459 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.143828 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.164672 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.184455 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.204931 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.224774 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.243851 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.265146 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.285257 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.304953 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.324921 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.344951 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.373834 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.384417 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.404230 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.424182 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.444058 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.463828 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.484574 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.504351 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.523795 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.544531 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.564379 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.584930 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.604235 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.625289 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.643915 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.664010 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.684925 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.704204 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.724976 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.744286 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.764741 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.785031 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.803830 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.825003 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.844854 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.865490 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.883929 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.904574 5018 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.927066 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.944595 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 14 06:52:04 crc kubenswrapper[5018]: I1014 06:52:04.964138 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.001843 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czt2r\" (UniqueName: \"kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r\") pod \"route-controller-manager-6576b87f9c-zq8rz\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.028498 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bcd\" (UniqueName: \"kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd\") pod \"oauth-openshift-558db77b4-htrtd\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.052103 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9mj9\" (UniqueName: \"kubernetes.io/projected/7cd77e7d-730d-4259-b784-4a7b8add26e0-kube-api-access-k9mj9\") pod \"machine-api-operator-5694c8668f-dp82w\" (UID: \"7cd77e7d-730d-4259-b784-4a7b8add26e0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.063525 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv9f9\" (UniqueName: \"kubernetes.io/projected/313736e9-309a-4784-a1b1-c7412d322eb5-kube-api-access-dv9f9\") pod \"openshift-config-operator-7777fb866f-f9c54\" (UID: \"313736e9-309a-4784-a1b1-c7412d322eb5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.067199 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.092994 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd47w\" (UniqueName: \"kubernetes.io/projected/c9a1d4cc-b532-44e2-ad14-d57f87be1bd4-kube-api-access-kd47w\") pod \"apiserver-76f77b778f-7dthh\" (UID: \"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4\") " pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.101110 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjg4c\" (UniqueName: \"kubernetes.io/projected/c6385ef6-72b4-49f4-b61b-546be9da202c-kube-api-access-wjg4c\") pod \"apiserver-7bbb656c7d-vf8lt\" (UID: \"c6385ef6-72b4-49f4-b61b-546be9da202c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.102164 5018 request.go:700] Waited for 1.928136875s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.128869 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.131370 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q7lm\" (UniqueName: \"kubernetes.io/projected/8241e590-8634-48c7-b07c-3ef44b4d2f85-kube-api-access-7q7lm\") pod \"openshift-apiserver-operator-796bbdcf4f-j6mpr\" (UID: \"8241e590-8634-48c7-b07c-3ef44b4d2f85\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.143476 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.153338 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.153886 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5mdx\" (UniqueName: \"kubernetes.io/projected/49c16b2d-7e27-4017-a41e-ab204084c2a7-kube-api-access-p5mdx\") pod \"cluster-samples-operator-665b6dd947-nvptz\" (UID: \"49c16b2d-7e27-4017-a41e-ab204084c2a7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.169028 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c12913c-de90-4dd3-a247-8a93e53565dc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nm2rb\" (UID: \"9c12913c-de90-4dd3-a247-8a93e53565dc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.194108 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qzsb\" (UniqueName: \"kubernetes.io/projected/98e1d068-3b1b-419d-9a5e-3617a8e93cc6-kube-api-access-4qzsb\") pod \"machine-approver-56656f9798-nd7jf\" (UID: \"98e1d068-3b1b-419d-9a5e-3617a8e93cc6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.206183 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f29c068e-309c-478a-a64d-8698c73028f2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rrm28\" (UID: \"f29c068e-309c-478a-a64d-8698c73028f2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.210846 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.236505 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.240964 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc8mn\" (UniqueName: \"kubernetes.io/projected/116c38b3-dbff-4f91-b057-93cc6ae4bb96-kube-api-access-kc8mn\") pod \"authentication-operator-69f744f599-p9sgv\" (UID: \"116c38b3-dbff-4f91-b057-93cc6ae4bb96\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.258019 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dt4l\" (UniqueName: \"kubernetes.io/projected/c01fd1f8-596f-445e-a209-c835c6d754f0-kube-api-access-8dt4l\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6x6l\" (UID: \"c01fd1f8-596f-445e-a209-c835c6d754f0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.261187 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.280916 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j26c\" (UniqueName: \"kubernetes.io/projected/d222c5fe-b248-425d-a810-83e03907f6a7-kube-api-access-8j26c\") pod \"console-operator-58897d9998-vc4k4\" (UID: \"d222c5fe-b248-425d-a810-83e03907f6a7\") " pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.298885 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8djll\" (UniqueName: \"kubernetes.io/projected/54e62cc6-54c6-4df2-b349-39bf04d702bc-kube-api-access-8djll\") pod \"downloads-7954f5f757-zvw8c\" (UID: \"54e62cc6-54c6-4df2-b349-39bf04d702bc\") " pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.304654 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.309844 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.323164 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.325698 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.329185 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.344019 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.346384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-545rn\" (UniqueName: \"kubernetes.io/projected/6559b957-5c56-410a-86d6-b75422bf82b3-kube-api-access-545rn\") pod \"router-default-5444994796-5hjlc\" (UID: \"6559b957-5c56-410a-86d6-b75422bf82b3\") " pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.357810 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.361597 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zxnx\" (UniqueName: \"kubernetes.io/projected/2245a631-4bde-4030-8318-2c7b57171224-kube-api-access-4zxnx\") pod \"cluster-image-registry-operator-dc59b4c8b-4m875\" (UID: \"2245a631-4bde-4030-8318-2c7b57171224\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.370274 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.379482 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvlq9\" (UniqueName: \"kubernetes.io/projected/68755a31-39a6-40e9-9317-1876c7393527-kube-api-access-zvlq9\") pod \"machine-config-operator-74547568cd-jsmbp\" (UID: \"68755a31-39a6-40e9-9317-1876c7393527\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.380499 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.390848 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" event={"ID":"17655b20-7a11-4fa4-a438-994408b65861","Type":"ContainerStarted","Data":"f0df6eef91a6ee36d93ef7154656b17ec7f1573506910a34acd3cceabcdcb4fd"} Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.408327 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f9c9584-72fd-4b26-98f9-917bb1978c7b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j764j\" (UID: \"6f9c9584-72fd-4b26-98f9-917bb1978c7b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.455358 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gd5x\" (UniqueName: \"kubernetes.io/projected/4dbe6d55-1450-4f4c-8d89-bb43c7429f05-kube-api-access-6gd5x\") pod \"dns-operator-744455d44c-jzjvj\" (UID: \"4dbe6d55-1450-4f4c-8d89-bb43c7429f05\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.460028 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljb8c\" (UniqueName: \"kubernetes.io/projected/e0f1ed6a-5861-471c-979b-76cfae275d2f-kube-api-access-ljb8c\") pod \"ingress-operator-5b745b69d9-sfkjm\" (UID: \"e0f1ed6a-5861-471c-979b-76cfae275d2f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.476222 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.483306 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f9c54"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.485139 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.495045 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrpkr\" (UniqueName: \"kubernetes.io/projected/ea6489fd-1b53-4ef0-9561-7691dd063b0c-kube-api-access-rrpkr\") pod \"olm-operator-6b444d44fb-4wlkr\" (UID: \"ea6489fd-1b53-4ef0-9561-7691dd063b0c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.498499 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506058 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfmpb\" (UniqueName: \"kubernetes.io/projected/01ed49d3-e7a8-48e4-9504-568e8d06494e-kube-api-access-tfmpb\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506093 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506114 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506141 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph9f2\" (UniqueName: \"kubernetes.io/projected/c844fe38-5193-422e-ac58-6bb43ee55180-kube-api-access-ph9f2\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506178 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506203 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506226 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4dg5\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506246 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506270 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506301 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506351 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttmf\" (UniqueName: \"kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506393 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506413 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-service-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506434 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506456 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-serving-cert\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506472 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gms6\" (UniqueName: \"kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506490 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506515 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506537 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506574 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-config\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506603 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-client\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506637 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506657 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506676 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506726 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-srv-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506785 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.506819 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.509481 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.009466821 +0000 UTC m=+142.593513448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.524224 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.525726 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.531351 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.532323 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.540870 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.546607 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" Oct 14 06:52:05 crc kubenswrapper[5018]: W1014 06:52:05.573488 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98e1d068_3b1b_419d_9a5e_3617a8e93cc6.slice/crio-94ba78d6787f17028b26c2092ba19deb339d736600e9246207058cd40e796c38 WatchSource:0}: Error finding container 94ba78d6787f17028b26c2092ba19deb339d736600e9246207058cd40e796c38: Status 404 returned error can't find the container with id 94ba78d6787f17028b26c2092ba19deb339d736600e9246207058cd40e796c38 Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611224 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611438 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611477 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611499 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103761a0-a528-43f2-ab80-4ad1bf882597-config\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.611722 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.111697955 +0000 UTC m=+142.695744652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611927 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-socket-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611965 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-csi-data-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.611984 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df392b5c-f08c-4085-a4b2-980775f79f9b-proxy-tls\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.612020 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bzr\" (UniqueName: \"kubernetes.io/projected/254036ee-39ae-4481-9398-3a520c7b4e1c-kube-api-access-d5bzr\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.612051 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt9kg\" (UniqueName: \"kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.612413 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.612997 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613030 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-registration-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613047 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613259 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfmpb\" (UniqueName: \"kubernetes.io/projected/01ed49d3-e7a8-48e4-9504-568e8d06494e-kube-api-access-tfmpb\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613307 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-key\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613389 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc4zm\" (UniqueName: \"kubernetes.io/projected/fc1de19a-1143-4a01-a51f-4c937eaa7e39-kube-api-access-zc4zm\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613449 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613471 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/65aea222-41ff-4806-ab92-ee7d19167a82-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613566 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613638 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613667 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph9f2\" (UniqueName: \"kubernetes.io/projected/c844fe38-5193-422e-ac58-6bb43ee55180-kube-api-access-ph9f2\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613692 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613724 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613745 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4dg5\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613770 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5sh5\" (UniqueName: \"kubernetes.io/projected/65aea222-41ff-4806-ab92-ee7d19167a82-kube-api-access-t5sh5\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613814 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614062 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.613843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614241 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614365 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-cabundle\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614414 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-node-bootstrap-token\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614433 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pggj4\" (UniqueName: \"kubernetes.io/projected/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-kube-api-access-pggj4\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614464 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614496 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttmf\" (UniqueName: \"kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614515 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqcx2\" (UniqueName: \"kubernetes.io/projected/7424d073-4b02-4d66-a69f-9f37fee2b473-kube-api-access-wqcx2\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614532 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614580 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-tmpfs\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614663 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614686 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lgft\" (UniqueName: \"kubernetes.io/projected/f81cde88-94b2-4d71-aaec-765256fe89d1-kube-api-access-8lgft\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614725 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-plugins-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614768 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614804 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-service-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614828 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbxtz\" (UniqueName: \"kubernetes.io/projected/103761a0-a528-43f2-ab80-4ad1bf882597-kube-api-access-wbxtz\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.614846 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-mountpoint-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615106 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1357a110-9b7b-4722-9985-31cb6140b37d-metrics-tls\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615174 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gms6\" (UniqueName: \"kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615214 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-serving-cert\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615235 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ff2h\" (UniqueName: \"kubernetes.io/projected/0e15820b-9507-497f-b17f-59cdae677e71-kube-api-access-2ff2h\") pod \"migrator-59844c95c7-vmv82\" (UID: \"0e15820b-9507-497f-b17f-59cdae677e71\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615253 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615271 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df392b5c-f08c-4085-a4b2-980775f79f9b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615342 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615394 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p6xn\" (UniqueName: \"kubernetes.io/projected/22b208b1-18a2-4e13-af40-f337838b8218-kube-api-access-7p6xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615457 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1357a110-9b7b-4722-9985-31cb6140b37d-config-volume\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615499 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615535 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615558 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-config\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615590 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615610 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-webhook-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615716 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22b208b1-18a2-4e13-af40-f337838b8218-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615820 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-client\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615873 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615891 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615906 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615945 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-certs\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615965 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.615981 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx2rf\" (UniqueName: \"kubernetes.io/projected/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-kube-api-access-gx2rf\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616043 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-srv-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616087 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616075 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616116 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103761a0-a528-43f2-ab80-4ad1bf882597-serving-cert\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616177 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc1de19a-1143-4a01-a51f-4c937eaa7e39-cert\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616196 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpr5q\" (UniqueName: \"kubernetes.io/projected/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-kube-api-access-qpr5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616255 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxvld\" (UniqueName: \"kubernetes.io/projected/df392b5c-f08c-4085-a4b2-980775f79f9b-kube-api-access-nxvld\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616314 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616353 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9pbs\" (UniqueName: \"kubernetes.io/projected/1357a110-9b7b-4722-9985-31cb6140b37d-kube-api-access-g9pbs\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.616374 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf8bc\" (UniqueName: \"kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.617291 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.617722 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.618823 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-service-ca\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.619172 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.619297 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.620553 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.620742 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.621705 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.121690335 +0000 UTC m=+142.705736962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.623356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ed49d3-e7a8-48e4-9504-568e8d06494e-config\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.631895 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.632348 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.635220 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.637093 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-srv-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.639434 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c844fe38-5193-422e-ac58-6bb43ee55180-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.640394 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.641256 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-serving-cert\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.641580 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.641938 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.642064 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.652088 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01ed49d3-e7a8-48e4-9504-568e8d06494e-etcd-client\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.652227 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfmpb\" (UniqueName: \"kubernetes.io/projected/01ed49d3-e7a8-48e4-9504-568e8d06494e-kube-api-access-tfmpb\") pod \"etcd-operator-b45778765-rgqhn\" (UID: \"01ed49d3-e7a8-48e4-9504-568e8d06494e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.660406 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph9f2\" (UniqueName: \"kubernetes.io/projected/c844fe38-5193-422e-ac58-6bb43ee55180-kube-api-access-ph9f2\") pod \"catalog-operator-68c6474976-6jf2w\" (UID: \"c844fe38-5193-422e-ac58-6bb43ee55180\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.681432 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttmf\" (UniqueName: \"kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf\") pod \"controller-manager-879f6c89f-6vg2h\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722323 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722582 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ff2h\" (UniqueName: \"kubernetes.io/projected/0e15820b-9507-497f-b17f-59cdae677e71-kube-api-access-2ff2h\") pod \"migrator-59844c95c7-vmv82\" (UID: \"0e15820b-9507-497f-b17f-59cdae677e71\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722678 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1357a110-9b7b-4722-9985-31cb6140b37d-metrics-tls\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722715 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722748 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p6xn\" (UniqueName: \"kubernetes.io/projected/22b208b1-18a2-4e13-af40-f337838b8218-kube-api-access-7p6xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722773 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df392b5c-f08c-4085-a4b2-980775f79f9b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722802 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1357a110-9b7b-4722-9985-31cb6140b37d-config-volume\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722831 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-webhook-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722852 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22b208b1-18a2-4e13-af40-f337838b8218-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722882 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722904 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-certs\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.722928 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx2rf\" (UniqueName: \"kubernetes.io/projected/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-kube-api-access-gx2rf\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723273 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103761a0-a528-43f2-ab80-4ad1bf882597-serving-cert\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723312 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc1de19a-1143-4a01-a51f-4c937eaa7e39-cert\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723361 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpr5q\" (UniqueName: \"kubernetes.io/projected/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-kube-api-access-qpr5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723383 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxvld\" (UniqueName: \"kubernetes.io/projected/df392b5c-f08c-4085-a4b2-980775f79f9b-kube-api-access-nxvld\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723432 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9pbs\" (UniqueName: \"kubernetes.io/projected/1357a110-9b7b-4722-9985-31cb6140b37d-kube-api-access-g9pbs\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723465 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf8bc\" (UniqueName: \"kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723521 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723547 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-socket-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723588 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-csi-data-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723649 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df392b5c-f08c-4085-a4b2-980775f79f9b-proxy-tls\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723717 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103761a0-a528-43f2-ab80-4ad1bf882597-config\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723746 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bzr\" (UniqueName: \"kubernetes.io/projected/254036ee-39ae-4481-9398-3a520c7b4e1c-kube-api-access-d5bzr\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723783 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt9kg\" (UniqueName: \"kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723822 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723847 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-registration-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723883 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723908 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-key\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723932 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc4zm\" (UniqueName: \"kubernetes.io/projected/fc1de19a-1143-4a01-a51f-4c937eaa7e39-kube-api-access-zc4zm\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723956 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/65aea222-41ff-4806-ab92-ee7d19167a82-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.723982 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724016 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5sh5\" (UniqueName: \"kubernetes.io/projected/65aea222-41ff-4806-ab92-ee7d19167a82-kube-api-access-t5sh5\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724056 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-cabundle\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724078 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-node-bootstrap-token\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724102 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqcx2\" (UniqueName: \"kubernetes.io/projected/7424d073-4b02-4d66-a69f-9f37fee2b473-kube-api-access-wqcx2\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724124 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pggj4\" (UniqueName: \"kubernetes.io/projected/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-kube-api-access-pggj4\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724145 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724165 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-tmpfs\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724212 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724246 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lgft\" (UniqueName: \"kubernetes.io/projected/f81cde88-94b2-4d71-aaec-765256fe89d1-kube-api-access-8lgft\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724270 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-plugins-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724311 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbxtz\" (UniqueName: \"kubernetes.io/projected/103761a0-a528-43f2-ab80-4ad1bf882597-kube-api-access-wbxtz\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724332 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-mountpoint-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.724936 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-csi-data-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.729704 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df392b5c-f08c-4085-a4b2-980775f79f9b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.731783 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1357a110-9b7b-4722-9985-31cb6140b37d-config-volume\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.733450 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.233423865 +0000 UTC m=+142.817470492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.734058 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.734189 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103761a0-a528-43f2-ab80-4ad1bf882597-config\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.735052 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-socket-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.737199 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.737820 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.738440 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-registration-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.739389 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-webhook-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.739527 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-plugins-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.739663 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-tmpfs\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.740325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-cabundle\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.741190 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gms6\" (UniqueName: \"kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6\") pod \"console-f9d7485db-sc99m\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.741457 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.741821 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f81cde88-94b2-4d71-aaec-765256fe89d1-mountpoint-dir\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.743473 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc1de19a-1143-4a01-a51f-4c937eaa7e39-cert\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.746127 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.748314 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.751114 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-certs\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.742254 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/254036ee-39ae-4481-9398-3a520c7b4e1c-signing-key\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.753071 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.754258 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.754910 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/65aea222-41ff-4806-ab92-ee7d19167a82-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.755281 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df392b5c-f08c-4085-a4b2-980775f79f9b-proxy-tls\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.756829 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7424d073-4b02-4d66-a69f-9f37fee2b473-node-bootstrap-token\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.756937 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22b208b1-18a2-4e13-af40-f337838b8218-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.759224 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.761048 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1357a110-9b7b-4722-9985-31cb6140b37d-metrics-tls\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.761748 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4dg5\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.762136 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.763673 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103761a0-a528-43f2-ab80-4ad1bf882597-serving-cert\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.793882 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.809479 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p6xn\" (UniqueName: \"kubernetes.io/projected/22b208b1-18a2-4e13-af40-f337838b8218-kube-api-access-7p6xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-j2fgv\" (UID: \"22b208b1-18a2-4e13-af40-f337838b8218\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.820954 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.822924 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx2rf\" (UniqueName: \"kubernetes.io/projected/3e8b4bd4-9f29-44f2-8b29-f619004ae69c-kube-api-access-gx2rf\") pod \"packageserver-d55dfcdfc-pzlrt\" (UID: \"3e8b4bd4-9f29-44f2-8b29-f619004ae69c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.826996 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.827691 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.327596775 +0000 UTC m=+142.911643402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.840548 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9pbs\" (UniqueName: \"kubernetes.io/projected/1357a110-9b7b-4722-9985-31cb6140b37d-kube-api-access-g9pbs\") pod \"dns-default-x2z8l\" (UID: \"1357a110-9b7b-4722-9985-31cb6140b37d\") " pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.855523 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.861642 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf8bc\" (UniqueName: \"kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc\") pod \"marketplace-operator-79b997595-mmmgm\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.879249 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-p9sgv"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.887817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5sh5\" (UniqueName: \"kubernetes.io/projected/65aea222-41ff-4806-ab92-ee7d19167a82-kube-api-access-t5sh5\") pod \"multus-admission-controller-857f4d67dd-8mzkh\" (UID: \"65aea222-41ff-4806-ab92-ee7d19167a82\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.890087 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.914374 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpr5q\" (UniqueName: \"kubernetes.io/projected/68d6bcd2-2ac8-410d-ae7d-d541314b9c3c-kube-api-access-qpr5q\") pod \"kube-storage-version-migrator-operator-b67b599dd-d259x\" (UID: \"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.928398 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:05 crc kubenswrapper[5018]: E1014 06:52:05.928943 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.428929063 +0000 UTC m=+143.012975690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.935186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ff2h\" (UniqueName: \"kubernetes.io/projected/0e15820b-9507-497f-b17f-59cdae677e71-kube-api-access-2ff2h\") pod \"migrator-59844c95c7-vmv82\" (UID: \"0e15820b-9507-497f-b17f-59cdae677e71\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.942006 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc4zm\" (UniqueName: \"kubernetes.io/projected/fc1de19a-1143-4a01-a51f-4c937eaa7e39-kube-api-access-zc4zm\") pod \"ingress-canary-hl8xt\" (UID: \"fc1de19a-1143-4a01-a51f-4c937eaa7e39\") " pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.950115 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.952331 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.956776 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.957745 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt9kg\" (UniqueName: \"kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg\") pod \"collect-profiles-29340405-2ndbp\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.968347 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.969305 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zvw8c"] Oct 14 06:52:05 crc kubenswrapper[5018]: I1014 06:52:05.997784 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.015304 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.015343 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.015841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bzr\" (UniqueName: \"kubernetes.io/projected/254036ee-39ae-4481-9398-3a520c7b4e1c-kube-api-access-d5bzr\") pod \"service-ca-9c57cc56f-dh52n\" (UID: \"254036ee-39ae-4481-9398-3a520c7b4e1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.018852 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxvld\" (UniqueName: \"kubernetes.io/projected/df392b5c-f08c-4085-a4b2-980775f79f9b-kube-api-access-nxvld\") pod \"machine-config-controller-84d6567774-r744g\" (UID: \"df392b5c-f08c-4085-a4b2-980775f79f9b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.032395 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.032776 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.532764954 +0000 UTC m=+143.116811581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.033300 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.034035 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pggj4\" (UniqueName: \"kubernetes.io/projected/3efff02c-77b3-4d8f-a64b-c4f0d7b778e1-kube-api-access-pggj4\") pod \"package-server-manager-789f6589d5-zfjfk\" (UID: \"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.035915 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.045187 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.046428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqcx2\" (UniqueName: \"kubernetes.io/projected/7424d073-4b02-4d66-a69f-9f37fee2b473-kube-api-access-wqcx2\") pod \"machine-config-server-mhqlx\" (UID: \"7424d073-4b02-4d66-a69f-9f37fee2b473\") " pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:06 crc kubenswrapper[5018]: W1014 06:52:06.050750 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod116c38b3_dbff_4f91_b057_93cc6ae4bb96.slice/crio-17f5daa7ac6e71778202a6567c3814b241648e5d027ec99b8fcbbd40c005b49c WatchSource:0}: Error finding container 17f5daa7ac6e71778202a6567c3814b241648e5d027ec99b8fcbbd40c005b49c: Status 404 returned error can't find the container with id 17f5daa7ac6e71778202a6567c3814b241648e5d027ec99b8fcbbd40c005b49c Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.051446 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hl8xt" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.056641 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mhqlx" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.069593 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lgft\" (UniqueName: \"kubernetes.io/projected/f81cde88-94b2-4d71-aaec-765256fe89d1-kube-api-access-8lgft\") pod \"csi-hostpathplugin-4j2t6\" (UID: \"f81cde88-94b2-4d71-aaec-765256fe89d1\") " pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.074469 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.075176 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.076864 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dp82w"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.077853 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.082656 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.084977 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbxtz\" (UniqueName: \"kubernetes.io/projected/103761a0-a528-43f2-ab80-4ad1bf882597-kube-api-access-wbxtz\") pod \"service-ca-operator-777779d784-wtvg8\" (UID: \"103761a0-a528-43f2-ab80-4ad1bf882597\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.136461 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.136826 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.63680678 +0000 UTC m=+143.220853407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.147897 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7dthh"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.226318 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.238802 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.239109 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.739098236 +0000 UTC m=+143.323144853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.257648 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.259987 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzjvj"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.293051 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.301090 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.310962 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.316528 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.320112 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.322216 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vc4k4"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.324416 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.340090 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.340463 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.840435974 +0000 UTC m=+143.424482601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.341028 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.342250 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.842232477 +0000 UTC m=+143.426279174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.443369 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.443979 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:06.943962436 +0000 UTC m=+143.528009063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.458983 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" event={"ID":"ea6489fd-1b53-4ef0-9561-7691dd063b0c","Type":"ContainerStarted","Data":"99c9a6f8906a44c843d640013aed65dff5538c06f2ab79103b4ca956f96c9aa7"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.468324 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" event={"ID":"7cd77e7d-730d-4259-b784-4a7b8add26e0","Type":"ContainerStarted","Data":"7beee12574fb8906a6ba0fe6110f4d3e76d2601c6c049028e289530f345a76c0"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.474261 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zvw8c" event={"ID":"54e62cc6-54c6-4df2-b349-39bf04d702bc","Type":"ContainerStarted","Data":"4042de9ef00756d19ed8fbff5e7cd05feb12ff972ebda45253a872c6f58e94a8"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.477779 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" event={"ID":"e0f1ed6a-5861-471c-979b-76cfae275d2f","Type":"ContainerStarted","Data":"76c4bd487be71ba126d72ef440680cb6e5f1e5fd9c7f8ac59a7a8e49d65c7ad1"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.480104 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mhqlx" event={"ID":"7424d073-4b02-4d66-a69f-9f37fee2b473","Type":"ContainerStarted","Data":"6414184a9ccfdbb9cc044966e57a8d3fea5eeb795f9e8bd515c71ee70455b0ea"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.486412 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" event={"ID":"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4","Type":"ContainerStarted","Data":"85105049eddfd50af773e104d2c45494be394f396fce259dbe72fd05680208af"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.493355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" event={"ID":"c01fd1f8-596f-445e-a209-c835c6d754f0","Type":"ContainerStarted","Data":"4441a206a2c4e76532b6e9d471aab90fb3268649a1c53868fc8ad7e30d8fbe18"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.545073 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" event={"ID":"4dbe6d55-1450-4f4c-8d89-bb43c7429f05","Type":"ContainerStarted","Data":"db38e4e9856a604756bd962136951969b9cb9adf7221d8cb94cfce3ae9e40624"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.553854 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.554164 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.054153811 +0000 UTC m=+143.638200438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.557528 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" event={"ID":"49c16b2d-7e27-4017-a41e-ab204084c2a7","Type":"ContainerStarted","Data":"a6cbb76e35d54841d64cc0edd68ea44ff28fd427881c4054b542d4104bf22914"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.570128 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.649180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" event={"ID":"17655b20-7a11-4fa4-a438-994408b65861","Type":"ContainerStarted","Data":"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.649216 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.654510 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.654995 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.154976733 +0000 UTC m=+143.739023360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.655067 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.655586 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.155578941 +0000 UTC m=+143.739625568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.671732 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" event={"ID":"8241e590-8634-48c7-b07c-3ef44b4d2f85","Type":"ContainerStarted","Data":"87fe8cbc2b99286ccace267656b2625aa77f1da0de3d31e6d03634c73ae12806"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.671791 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" event={"ID":"8241e590-8634-48c7-b07c-3ef44b4d2f85","Type":"ContainerStarted","Data":"2dbfd20e14a6df2a8363b04d3956772c5b875a2bc56697ae040762a8a473b0db"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.675764 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" event={"ID":"c6385ef6-72b4-49f4-b61b-546be9da202c","Type":"ContainerStarted","Data":"d7f372ec3074f3c4f6179c9940f005a6208e0ca54e572b2cfd48fa41d8f9267c"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.679429 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" event={"ID":"f29c068e-309c-478a-a64d-8698c73028f2","Type":"ContainerStarted","Data":"37a0e6990bb817b1030d3041bd685d7619668318dced3537e2257be228fcbae0"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.681129 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" event={"ID":"98e1d068-3b1b-419d-9a5e-3617a8e93cc6","Type":"ContainerStarted","Data":"193dc8906585ff4494c4bcba5ba4f8e0195c57021a7f91438b96d3b9e1479841"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.681154 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" event={"ID":"98e1d068-3b1b-419d-9a5e-3617a8e93cc6","Type":"ContainerStarted","Data":"94ba78d6787f17028b26c2092ba19deb339d736600e9246207058cd40e796c38"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.681998 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" event={"ID":"9c12913c-de90-4dd3-a247-8a93e53565dc","Type":"ContainerStarted","Data":"2da9816fd1b0963cb223b3343d0551026e5a4725319ecc52932d1b718da427d0"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.686422 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" event={"ID":"aaa522ed-38fb-40f0-9015-b97956137bc8","Type":"ContainerStarted","Data":"0372713277d9b65db71add3414ab56be57f6e7f86199591b8c0dfe6862dba6c5"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.690240 5018 generic.go:334] "Generic (PLEG): container finished" podID="313736e9-309a-4784-a1b1-c7412d322eb5" containerID="acf0cbd58beab1de807daeaec1a41d7fc7fc920cf5afaf773c1f606b74868448" exitCode=0 Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.690810 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" event={"ID":"313736e9-309a-4784-a1b1-c7412d322eb5","Type":"ContainerDied","Data":"acf0cbd58beab1de807daeaec1a41d7fc7fc920cf5afaf773c1f606b74868448"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.690855 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" event={"ID":"313736e9-309a-4784-a1b1-c7412d322eb5","Type":"ContainerStarted","Data":"f47038bd2e6f6e3cd88bac351ce5bcd15e45530cdb5730e350b8eb0912d28b26"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.693313 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5hjlc" event={"ID":"6559b957-5c56-410a-86d6-b75422bf82b3","Type":"ContainerStarted","Data":"f73093fda4ce9983683a2f1d327cf6bb1143b71212e09e50f2da51d0a764164b"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.693342 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5hjlc" event={"ID":"6559b957-5c56-410a-86d6-b75422bf82b3","Type":"ContainerStarted","Data":"0fd909cc3a49e8a5573317d40135003be44109ff50755ea384b98c63aa5eae1a"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.696422 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" event={"ID":"116c38b3-dbff-4f91-b057-93cc6ae4bb96","Type":"ContainerStarted","Data":"17f5daa7ac6e71778202a6567c3814b241648e5d027ec99b8fcbbd40c005b49c"} Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.753556 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.755811 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.757064 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.257048243 +0000 UTC m=+143.841094870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.759231 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.761168 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.772130 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt"] Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.858295 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.858554 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.358543256 +0000 UTC m=+143.942589883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: W1014 06:52:06.891685 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83a99049_d48d_4e15_bb4d_3eb09e9eb971.slice/crio-4cf80d3661af2a3d6001824a3ade4db10d002607f7169d20ec92f88323133f75 WatchSource:0}: Error finding container 4cf80d3661af2a3d6001824a3ade4db10d002607f7169d20ec92f88323133f75: Status 404 returned error can't find the container with id 4cf80d3661af2a3d6001824a3ade4db10d002607f7169d20ec92f88323133f75 Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.919888 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" podStartSLOduration=121.919875664 podStartE2EDuration="2m1.919875664s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:06.917889496 +0000 UTC m=+143.501936123" watchObservedRunningTime="2025-10-14 06:52:06.919875664 +0000 UTC m=+143.503922361" Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.963714 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.963895 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.46387459 +0000 UTC m=+144.047921217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:06 crc kubenswrapper[5018]: I1014 06:52:06.963963 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:06 crc kubenswrapper[5018]: E1014 06:52:06.964273 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.464263981 +0000 UTC m=+144.048310608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.065170 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.065338 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.565315831 +0000 UTC m=+144.149362458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.065538 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.065931 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.565916368 +0000 UTC m=+144.149962995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.167243 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.168074 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.667950047 +0000 UTC m=+144.251996684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.178360 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.180003 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.180728 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.680711967 +0000 UTC m=+144.264758604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.247972 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.258744 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.262713 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.276335 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6mpr" podStartSLOduration=123.276313669 podStartE2EDuration="2m3.276313669s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:07.218587495 +0000 UTC m=+143.802634122" watchObservedRunningTime="2025-10-14 06:52:07.276313669 +0000 UTC m=+143.860360296" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.283144 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.283532 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.783512247 +0000 UTC m=+144.367558884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.296982 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rgqhn"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.305916 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.312785 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4j2t6"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.315695 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8mzkh"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.318481 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5hjlc" podStartSLOduration=122.318466811 podStartE2EDuration="2m2.318466811s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:07.263740254 +0000 UTC m=+143.847786881" watchObservedRunningTime="2025-10-14 06:52:07.318466811 +0000 UTC m=+143.902513438" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.322515 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dh52n"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.353245 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x2z8l"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.358148 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.369784 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hl8xt"] Oct 14 06:52:07 crc kubenswrapper[5018]: W1014 06:52:07.376257 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01ed49d3_e7a8_48e4_9504_568e8d06494e.slice/crio-493db053d1266de336e018aeac2a8f167a599f5ed459d8bd73c50de56353385a WatchSource:0}: Error finding container 493db053d1266de336e018aeac2a8f167a599f5ed459d8bd73c50de56353385a: Status 404 returned error can't find the container with id 493db053d1266de336e018aeac2a8f167a599f5ed459d8bd73c50de56353385a Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.376591 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r744g"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.387292 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.387573 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.887563554 +0000 UTC m=+144.471610171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: W1014 06:52:07.408466 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65aea222_41ff_4806_ab92_ee7d19167a82.slice/crio-77725e2c1c8b553cb2c28f8ac09676674f87a5ec68ca0e502690b662d92964d5 WatchSource:0}: Error finding container 77725e2c1c8b553cb2c28f8ac09676674f87a5ec68ca0e502690b662d92964d5: Status 404 returned error can't find the container with id 77725e2c1c8b553cb2c28f8ac09676674f87a5ec68ca0e502690b662d92964d5 Oct 14 06:52:07 crc kubenswrapper[5018]: W1014 06:52:07.419806 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod254036ee_39ae_4481_9398_3a520c7b4e1c.slice/crio-18c5d9c78319a92f79d3289ffd6b01864d39a440ee99de28a17e5280bc8f2227 WatchSource:0}: Error finding container 18c5d9c78319a92f79d3289ffd6b01864d39a440ee99de28a17e5280bc8f2227: Status 404 returned error can't find the container with id 18c5d9c78319a92f79d3289ffd6b01864d39a440ee99de28a17e5280bc8f2227 Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.439747 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8"] Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.487746 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.488496 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:07.988461 +0000 UTC m=+144.572507627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.534959 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.547892 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:07 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:07 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:07 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.548246 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.589978 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.590245 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.09023454 +0000 UTC m=+144.674281167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.691386 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.715377 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.215349768 +0000 UTC m=+144.799396405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.721324 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zvw8c" event={"ID":"54e62cc6-54c6-4df2-b349-39bf04d702bc","Type":"ContainerStarted","Data":"506992278e14aa630878a61a73c65dfb21e3ecd38624ae40d508bd092eccd48f"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.731607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" event={"ID":"c844fe38-5193-422e-ac58-6bb43ee55180","Type":"ContainerStarted","Data":"fc360ff94ebbfbad4a78fef2fd5cf10f69fdecacac86b857385cac80be591cc0"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.738395 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.743585 5018 patch_prober.go:28] interesting pod/downloads-7954f5f757-zvw8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.743666 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zvw8c" podUID="54e62cc6-54c6-4df2-b349-39bf04d702bc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.746856 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" event={"ID":"254036ee-39ae-4481-9398-3a520c7b4e1c","Type":"ContainerStarted","Data":"18c5d9c78319a92f79d3289ffd6b01864d39a440ee99de28a17e5280bc8f2227"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.750346 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" event={"ID":"f81cde88-94b2-4d71-aaec-765256fe89d1","Type":"ContainerStarted","Data":"16ade9cdd6f5bf5134a9e875183b63db5b67506aac99d0911bdb9b434faf09b9"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.751444 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" event={"ID":"df392b5c-f08c-4085-a4b2-980775f79f9b","Type":"ContainerStarted","Data":"ab7037755d5c2677346e58957ca1fa2ce51528a527845608e5cf2ffbc7b2e186"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.752999 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" event={"ID":"9c12913c-de90-4dd3-a247-8a93e53565dc","Type":"ContainerStarted","Data":"1ab262894933155dc816b6488609e985cf3f6aaf87c70929ed11aa6af7c45010"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.758192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mhqlx" event={"ID":"7424d073-4b02-4d66-a69f-9f37fee2b473","Type":"ContainerStarted","Data":"b84a2516928bfe1f55d5b63d0a53468b183d70543f2b17e13856c46cc379b439"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.761174 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" event={"ID":"b7a8f667-d3bf-43eb-b046-291b70da5c13","Type":"ContainerStarted","Data":"aee0f9ad67fd7bc4b2ce9d81ecff8b0e9d08769f48992a68c240e56e83ed9e02"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.769396 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" event={"ID":"103761a0-a528-43f2-ab80-4ad1bf882597","Type":"ContainerStarted","Data":"1c6bcebb0a99e388a83ed16f08a11554a8f821deb41f69ed61bcd4b4a0127785"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.780598 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" event={"ID":"7cd77e7d-730d-4259-b784-4a7b8add26e0","Type":"ContainerStarted","Data":"1be566d0e21a97a44d6af74905959d551f020832e92ff71a3f2e433282df8e9f"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.782583 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" event={"ID":"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c","Type":"ContainerStarted","Data":"6e835b69666ba7ddf96b500f2fc100665ce85a76562ab16e5287a77e3cbb601d"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.784916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" event={"ID":"4dbe6d55-1450-4f4c-8d89-bb43c7429f05","Type":"ContainerStarted","Data":"73e757d5265be509de4cc3079c33c1b4144efff33f867b3cb929f27aeba7d799"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.789041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hl8xt" event={"ID":"fc1de19a-1143-4a01-a51f-4c937eaa7e39","Type":"ContainerStarted","Data":"c62047a856b05ba7ef5de715c0d9011af378aaff098008493e3529f35556ccc3"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.794498 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" event={"ID":"6f9c9584-72fd-4b26-98f9-917bb1978c7b","Type":"ContainerStarted","Data":"0fd1527ba7c743da1136de6df0a0b4703bfc7549029b40b8b69a678f374b8d54"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.794568 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" event={"ID":"6f9c9584-72fd-4b26-98f9-917bb1978c7b","Type":"ContainerStarted","Data":"3f8de280ceb521377612db29a8a216dbac03ff236c524253ddbafd5dcb3d10e9"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.795147 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.795506 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.295490192 +0000 UTC m=+144.879536819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.799024 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" event={"ID":"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1","Type":"ContainerStarted","Data":"c55cf3ee2f99f13f21dfc46c48598d54e7870f3bdd93addf86b14d44246cf5bd"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.802350 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" event={"ID":"22b208b1-18a2-4e13-af40-f337838b8218","Type":"ContainerStarted","Data":"c948430aa0d61db5abe902d05a66d339ec1b9836764582f985335c787c0ca163"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.807574 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" event={"ID":"68755a31-39a6-40e9-9317-1876c7393527","Type":"ContainerStarted","Data":"2cad4849e64b4dc2f1e64ff92b0f9901c7d93091595cf14ca0495bbd0ab75de7"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.807628 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" event={"ID":"68755a31-39a6-40e9-9317-1876c7393527","Type":"ContainerStarted","Data":"515f627e03920b8a94ce633b427b053b7ae159301770bcb3e6d42341f0105282"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.808762 5018 generic.go:334] "Generic (PLEG): container finished" podID="c9a1d4cc-b532-44e2-ad14-d57f87be1bd4" containerID="f2012fc952019729a9225334da65aacadc018da9ccf2dcdba1b7220164b88ec0" exitCode=0 Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.808892 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" event={"ID":"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4","Type":"ContainerDied","Data":"f2012fc952019729a9225334da65aacadc018da9ccf2dcdba1b7220164b88ec0"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.810916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2z8l" event={"ID":"1357a110-9b7b-4722-9985-31cb6140b37d","Type":"ContainerStarted","Data":"66d5685a422382c21cfc10186a47de5e2ddf340436eba4b5fca3b1050631bc8f"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.812743 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" event={"ID":"49c16b2d-7e27-4017-a41e-ab204084c2a7","Type":"ContainerStarted","Data":"5d58e9b0165dd181fc697835236e7afa32e91b84c4c8e2a97efd2f29964771e5"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.828316 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" event={"ID":"01ed49d3-e7a8-48e4-9504-568e8d06494e","Type":"ContainerStarted","Data":"493db053d1266de336e018aeac2a8f167a599f5ed459d8bd73c50de56353385a"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.830514 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" event={"ID":"f29c068e-309c-478a-a64d-8698c73028f2","Type":"ContainerStarted","Data":"ed4d2e1af2c768aff6b4641b724d5798244768209b5336907ad2675d6358de50"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.836297 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" event={"ID":"aaa522ed-38fb-40f0-9015-b97956137bc8","Type":"ContainerStarted","Data":"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.837107 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.839116 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" event={"ID":"c01fd1f8-596f-445e-a209-c835c6d754f0","Type":"ContainerStarted","Data":"07605254c5cbf6f55e4f81b575fc8d3ba970ab651103364193691f95e53796d6"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.842944 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" event={"ID":"98e1d068-3b1b-419d-9a5e-3617a8e93cc6","Type":"ContainerStarted","Data":"342f1e152233970bbcb82a8ea310e260ca16a5efaff1c18919717cbce8660ede"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.845148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" event={"ID":"0e15820b-9507-497f-b17f-59cdae677e71","Type":"ContainerStarted","Data":"69d2729e6f34c9ba8160ffc570f514af20485fbe69a58be0a70e05a365a51cfa"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.896019 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sc99m" event={"ID":"83a99049-d48d-4e15-bb4d-3eb09e9eb971","Type":"ContainerStarted","Data":"4cf80d3661af2a3d6001824a3ade4db10d002607f7169d20ec92f88323133f75"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.896228 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:07 crc kubenswrapper[5018]: E1014 06:52:07.897537 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.39751777 +0000 UTC m=+144.981564397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.927847 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" event={"ID":"116c38b3-dbff-4f91-b057-93cc6ae4bb96","Type":"ContainerStarted","Data":"dd30e227df7a08f7f078297717c634e6854435dc4043ff901d5735cf357573a9"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.946552 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" event={"ID":"e0f1ed6a-5861-471c-979b-76cfae275d2f","Type":"ContainerStarted","Data":"2a15d9e74b86ba825979f9c79b3cb2b2ddaf72fbe8c4950a3f8fc27be7c30e13"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.954708 5018 generic.go:334] "Generic (PLEG): container finished" podID="c6385ef6-72b4-49f4-b61b-546be9da202c" containerID="70cbcd94830ca5e2b25b2f7112460bbf18af2be1fe021b3421699e56df7a99b4" exitCode=0 Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.954791 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" event={"ID":"c6385ef6-72b4-49f4-b61b-546be9da202c","Type":"ContainerDied","Data":"70cbcd94830ca5e2b25b2f7112460bbf18af2be1fe021b3421699e56df7a99b4"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.965418 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" event={"ID":"d222c5fe-b248-425d-a810-83e03907f6a7","Type":"ContainerStarted","Data":"c2013896574626bd16446ccc69f3fbf844003680e6a9ccbfc3f360b2f9b048f3"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.965728 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" event={"ID":"d222c5fe-b248-425d-a810-83e03907f6a7","Type":"ContainerStarted","Data":"7a29e236da702db4e7c2ad5695ad35b4b9e9d40f143cba842fe1bbdc2c0b0608"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.966677 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.969510 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" event={"ID":"ea6489fd-1b53-4ef0-9561-7691dd063b0c","Type":"ContainerStarted","Data":"06966cae77143ffb8363f739a898d7de63844edf1f301a7d8f987963085435e8"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.969996 5018 patch_prober.go:28] interesting pod/console-operator-58897d9998-vc4k4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.970036 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" podUID="d222c5fe-b248-425d-a810-83e03907f6a7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.970388 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.970816 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" event={"ID":"65aea222-41ff-4806-ab92-ee7d19167a82","Type":"ContainerStarted","Data":"77725e2c1c8b553cb2c28f8ac09676674f87a5ec68ca0e502690b662d92964d5"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.971753 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" event={"ID":"3e8b4bd4-9f29-44f2-8b29-f619004ae69c","Type":"ContainerStarted","Data":"ef0ed1bb6323f87ce260089fe457c4f57e25d98dfb6cbaa19ce7a9772681bd77"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.980837 5018 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4wlkr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.980874 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" podUID="ea6489fd-1b53-4ef0-9561-7691dd063b0c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.983167 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" event={"ID":"2245a631-4bde-4030-8318-2c7b57171224","Type":"ContainerStarted","Data":"1fb44f9edc474b26a664c9ad13d81d9b3e6b955ddb1929a625106e7b638514d8"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.983192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" event={"ID":"2245a631-4bde-4030-8318-2c7b57171224","Type":"ContainerStarted","Data":"8f48f4509bb666b9523a499bcb924f0d3b3b6b14811ee9703b18924b718a7b9c"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.985467 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" event={"ID":"313736e9-309a-4784-a1b1-c7412d322eb5","Type":"ContainerStarted","Data":"35d6038202f5bbe77590e3b7da6c58a57e94ba098f50faa6fded790aa33b45d1"} Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.985814 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:07 crc kubenswrapper[5018]: I1014 06:52:07.997128 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" event={"ID":"b03f9bd8-9597-4358-aae3-4d3cceaf3a47","Type":"ContainerStarted","Data":"fd4fe0158ddeec38f260290b3845cc5576711b6b4178c3784effd7ccc9c8958c"} Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.003449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" event={"ID":"ac8f3d02-ac5f-4879-a8d3-66c226c42964","Type":"ContainerStarted","Data":"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930"} Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.003492 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" event={"ID":"ac8f3d02-ac5f-4879-a8d3-66c226c42964","Type":"ContainerStarted","Data":"445fc5522c9dfdc02307597880bfac99d795c74ac8a92db8d6a71819f843d6ab"} Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.003519 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.003624 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.005524 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.505508111 +0000 UTC m=+145.089554738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.032266 5018 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6vg2h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.032312 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.107374 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.108587 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.608568769 +0000 UTC m=+145.192615406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.209212 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.209550 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.709537917 +0000 UTC m=+145.293584544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.293995 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rrm28" podStartSLOduration=123.293971515 podStartE2EDuration="2m3.293971515s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.289189586 +0000 UTC m=+144.873236233" watchObservedRunningTime="2025-10-14 06:52:08.293971515 +0000 UTC m=+144.878018152" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.294507 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" podStartSLOduration=123.2945022 podStartE2EDuration="2m3.2945022s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.244640254 +0000 UTC m=+144.828686881" watchObservedRunningTime="2025-10-14 06:52:08.2945022 +0000 UTC m=+144.878548827" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.311049 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.360425 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.860402231 +0000 UTC m=+145.444448858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.401528 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" podStartSLOduration=124.401511073 podStartE2EDuration="2m4.401511073s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.399791203 +0000 UTC m=+144.983837830" watchObservedRunningTime="2025-10-14 06:52:08.401511073 +0000 UTC m=+144.985557700" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.404359 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" podStartSLOduration=123.404349045 podStartE2EDuration="2m3.404349045s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.360643368 +0000 UTC m=+144.944689995" watchObservedRunningTime="2025-10-14 06:52:08.404349045 +0000 UTC m=+144.988395662" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.428901 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.429437 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:08.929425712 +0000 UTC m=+145.513472339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.440310 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4m875" podStartSLOduration=123.440295037 podStartE2EDuration="2m3.440295037s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.438559697 +0000 UTC m=+145.022606324" watchObservedRunningTime="2025-10-14 06:52:08.440295037 +0000 UTC m=+145.024341664" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.490667 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6x6l" podStartSLOduration=123.490648377 podStartE2EDuration="2m3.490648377s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.489901106 +0000 UTC m=+145.073947743" watchObservedRunningTime="2025-10-14 06:52:08.490648377 +0000 UTC m=+145.074695004" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.530013 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.530294 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.030277406 +0000 UTC m=+145.614324033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.530433 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.530761 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.03075391 +0000 UTC m=+145.614800537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.539552 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:08 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:08 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:08 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.540118 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.551702 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" podStartSLOduration=123.551684587 podStartE2EDuration="2m3.551684587s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.53556504 +0000 UTC m=+145.119611667" watchObservedRunningTime="2025-10-14 06:52:08.551684587 +0000 UTC m=+145.135731214" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.614569 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" podStartSLOduration=123.61455534 podStartE2EDuration="2m3.61455534s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.614055685 +0000 UTC m=+145.198102312" watchObservedRunningTime="2025-10-14 06:52:08.61455534 +0000 UTC m=+145.198601957" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.616400 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nm2rb" podStartSLOduration=123.616390793 podStartE2EDuration="2m3.616390793s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.581382708 +0000 UTC m=+145.165429325" watchObservedRunningTime="2025-10-14 06:52:08.616390793 +0000 UTC m=+145.200437420" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.633293 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.634076 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.134051875 +0000 UTC m=+145.718098492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.652218 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-zvw8c" podStartSLOduration=123.652199211 podStartE2EDuration="2m3.652199211s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.649100682 +0000 UTC m=+145.233147309" watchObservedRunningTime="2025-10-14 06:52:08.652199211 +0000 UTC m=+145.236245838" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.713962 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-p9sgv" podStartSLOduration=124.713945392 podStartE2EDuration="2m4.713945392s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.68596128 +0000 UTC m=+145.270007917" watchObservedRunningTime="2025-10-14 06:52:08.713945392 +0000 UTC m=+145.297992019" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.718025 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j764j" podStartSLOduration=123.71801412 podStartE2EDuration="2m3.71801412s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.716942509 +0000 UTC m=+145.300989136" watchObservedRunningTime="2025-10-14 06:52:08.71801412 +0000 UTC m=+145.302060747" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.736030 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.736400 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.236384092 +0000 UTC m=+145.820430719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.743539 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.771735 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nd7jf" podStartSLOduration=124.771691376 podStartE2EDuration="2m4.771691376s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.771166921 +0000 UTC m=+145.355213558" watchObservedRunningTime="2025-10-14 06:52:08.771691376 +0000 UTC m=+145.355738003" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.838053 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.838702 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.338686998 +0000 UTC m=+145.922733625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.847533 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" podStartSLOduration=123.847516555 podStartE2EDuration="2m3.847516555s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.84253529 +0000 UTC m=+145.426581927" watchObservedRunningTime="2025-10-14 06:52:08.847516555 +0000 UTC m=+145.431563182" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.922247 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-mhqlx" podStartSLOduration=6.922221131 podStartE2EDuration="6.922221131s" podCreationTimestamp="2025-10-14 06:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.881870681 +0000 UTC m=+145.465917308" watchObservedRunningTime="2025-10-14 06:52:08.922221131 +0000 UTC m=+145.506267758" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.923458 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" podStartSLOduration=123.923446986 podStartE2EDuration="2m3.923446986s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:08.918126822 +0000 UTC m=+145.502173449" watchObservedRunningTime="2025-10-14 06:52:08.923446986 +0000 UTC m=+145.507493613" Oct 14 06:52:08 crc kubenswrapper[5018]: I1014 06:52:08.942664 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:08 crc kubenswrapper[5018]: E1014 06:52:08.943088 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.443070195 +0000 UTC m=+146.027116822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.029959 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" event={"ID":"22b208b1-18a2-4e13-af40-f337838b8218","Type":"ContainerStarted","Data":"233714f6d5c41c2c544c7361bd5ec8a1edf1e9aa02a1b409bb3dbccdde83a0f2"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.045112 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-nvptz" event={"ID":"49c16b2d-7e27-4017-a41e-ab204084c2a7","Type":"ContainerStarted","Data":"c2f6c9715e7cf72d56de3ae1d9f5e7df71478fca7b9f86e21bbdad6dff463da8"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.046255 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.046767 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.546747251 +0000 UTC m=+146.130793878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.051205 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j2fgv" podStartSLOduration=124.05118964 podStartE2EDuration="2m4.05118964s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.048385589 +0000 UTC m=+145.632432216" watchObservedRunningTime="2025-10-14 06:52:09.05118964 +0000 UTC m=+145.635236267" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.062295 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sc99m" event={"ID":"83a99049-d48d-4e15-bb4d-3eb09e9eb971","Type":"ContainerStarted","Data":"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.069043 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dp82w" event={"ID":"7cd77e7d-730d-4259-b784-4a7b8add26e0","Type":"ContainerStarted","Data":"2b127e969941673da878e80ee548914dbf7ed335e143c92742bba0a984c4c158"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.072913 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" event={"ID":"3e8b4bd4-9f29-44f2-8b29-f619004ae69c","Type":"ContainerStarted","Data":"e293440ae47a32e88081eb3505d63c1a86612e5a99abf69243a67e73dccd611e"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.072956 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.081357 5018 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pzlrt container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.081422 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" podUID="3e8b4bd4-9f29-44f2-8b29-f619004ae69c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.084481 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hl8xt" event={"ID":"fc1de19a-1143-4a01-a51f-4c937eaa7e39","Type":"ContainerStarted","Data":"473cb006d9f8c084642cc617de4e43f3e37c3ba867034d5db04eeaf78f7b3569"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.090465 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" event={"ID":"103761a0-a528-43f2-ab80-4ad1bf882597","Type":"ContainerStarted","Data":"ded8c6eb21dbdd0a0c0a62f0bd2d28349fdf9946a65241576063dbcccd6a9e3b"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.097214 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2z8l" event={"ID":"1357a110-9b7b-4722-9985-31cb6140b37d","Type":"ContainerStarted","Data":"aa01fa0ee7ff6b2b41b4f7481f2c62c10baa2d9741f84c3de4a22b05d02f7e17"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.105513 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sc99m" podStartSLOduration=124.105497494 podStartE2EDuration="2m4.105497494s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.097385209 +0000 UTC m=+145.681431836" watchObservedRunningTime="2025-10-14 06:52:09.105497494 +0000 UTC m=+145.689544121" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.115812 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" event={"ID":"68d6bcd2-2ac8-410d-ae7d-d541314b9c3c","Type":"ContainerStarted","Data":"6926a7dc9f67d1c9843aad8ffbdb99b895359701e2f545fe14b727fcdbbe0acb"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.150496 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.156373 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" podStartSLOduration=124.156350739 podStartE2EDuration="2m4.156350739s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.130595202 +0000 UTC m=+145.714641829" watchObservedRunningTime="2025-10-14 06:52:09.156350739 +0000 UTC m=+145.740397376" Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.161710 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.661685354 +0000 UTC m=+146.245731981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.181274 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" event={"ID":"01ed49d3-e7a8-48e4-9504-568e8d06494e","Type":"ContainerStarted","Data":"18232ec7d9441dfb39df077eaba730f575921f9d1c4224e461192023131003f4"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.191514 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" event={"ID":"65aea222-41ff-4806-ab92-ee7d19167a82","Type":"ContainerStarted","Data":"bef90133e7ee6e96ccb0b13cac81782dc9588bfe3813e1b84b7ab82b4073f13e"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.196309 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wtvg8" podStartSLOduration=124.196290517 podStartE2EDuration="2m4.196290517s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.191994452 +0000 UTC m=+145.776041089" watchObservedRunningTime="2025-10-14 06:52:09.196290517 +0000 UTC m=+145.780337144" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.211867 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hl8xt" podStartSLOduration=7.211853278 podStartE2EDuration="7.211853278s" podCreationTimestamp="2025-10-14 06:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.209662855 +0000 UTC m=+145.793709482" watchObservedRunningTime="2025-10-14 06:52:09.211853278 +0000 UTC m=+145.795899895" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.236926 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" event={"ID":"c844fe38-5193-422e-ac58-6bb43ee55180","Type":"ContainerStarted","Data":"dd36967c6d6a4344240037e7f7eafc5c5476dbcea70f68202437a9c65f856104"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.237820 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.239102 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rgqhn" podStartSLOduration=124.239092838 podStartE2EDuration="2m4.239092838s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.2381429 +0000 UTC m=+145.822189527" watchObservedRunningTime="2025-10-14 06:52:09.239092838 +0000 UTC m=+145.823139465" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.244634 5018 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6jf2w container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.244676 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" podUID="c844fe38-5193-422e-ac58-6bb43ee55180" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.249439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" event={"ID":"e0f1ed6a-5861-471c-979b-76cfae275d2f","Type":"ContainerStarted","Data":"f71995f70b11b073f8f552c7a3cea5a1bf998f21c3f3ae8e6c0d43f5abba8056"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.251609 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.252750 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.752730703 +0000 UTC m=+146.336777330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.266959 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d259x" podStartSLOduration=124.266937715 podStartE2EDuration="2m4.266937715s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.256174473 +0000 UTC m=+145.840221100" watchObservedRunningTime="2025-10-14 06:52:09.266937715 +0000 UTC m=+145.850984342" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.288327 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" event={"ID":"68755a31-39a6-40e9-9317-1876c7393527","Type":"ContainerStarted","Data":"611e889754f404a7b3b5ceb2e2bafcddac8c10e7ed0ad7edf000eba504b1ced5"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.304654 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" event={"ID":"b7a8f667-d3bf-43eb-b046-291b70da5c13","Type":"ContainerStarted","Data":"24c829c769ae1e7e691b4788880cac2bc276230b80d85df18f092f5b48c7ba7d"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.317892 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" event={"ID":"b03f9bd8-9597-4358-aae3-4d3cceaf3a47","Type":"ContainerStarted","Data":"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.318274 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.320995 5018 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mmmgm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.321154 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.321553 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" podStartSLOduration=124.321539598 podStartE2EDuration="2m4.321539598s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.318436278 +0000 UTC m=+145.902482905" watchObservedRunningTime="2025-10-14 06:52:09.321539598 +0000 UTC m=+145.905586225" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.326140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" event={"ID":"df392b5c-f08c-4085-a4b2-980775f79f9b","Type":"ContainerStarted","Data":"ab54ad024e4b3893c2e03ce12130690c64cbdabe14d65650f145096fdeef5fac"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.326195 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" event={"ID":"df392b5c-f08c-4085-a4b2-980775f79f9b","Type":"ContainerStarted","Data":"6b7fdc69f773a16ee4d9e160a9f814a8ea8dc355e5f5719a8e2a46cb5dc39863"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.336309 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfkjm" podStartSLOduration=124.336287626 podStartE2EDuration="2m4.336287626s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.29709636 +0000 UTC m=+145.881142997" watchObservedRunningTime="2025-10-14 06:52:09.336287626 +0000 UTC m=+145.920334253" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.341009 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" event={"ID":"254036ee-39ae-4481-9398-3a520c7b4e1c","Type":"ContainerStarted","Data":"f7b14458f303fab0616dea1476adcf5dd89d3e1163379392dad378618d7b9073"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.348889 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" event={"ID":"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1","Type":"ContainerStarted","Data":"9abfa16c0f1effaf4a09455d8ce3a06e187aab415fde954499be1fa2c3d96b39"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.348933 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" event={"ID":"3efff02c-77b3-4d8f-a64b-c4f0d7b778e1","Type":"ContainerStarted","Data":"97c2d169792c8772bc923b22212a2bc39dc664e16d1ab419bdbaf602180a9d76"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.349448 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.354785 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" event={"ID":"0e15820b-9507-497f-b17f-59cdae677e71","Type":"ContainerStarted","Data":"18858fa149896a63350fe437530b7affaa7d92ccdb2781ab243b198f6d8a165d"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.354820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" event={"ID":"0e15820b-9507-497f-b17f-59cdae677e71","Type":"ContainerStarted","Data":"14b9a9ef03c24c802a1d615e11a187937c2b7cc76c2abba5d500337e5949df89"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.356040 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.356362 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.856350688 +0000 UTC m=+146.440397305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.358649 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" podStartSLOduration=124.358636644 podStartE2EDuration="2m4.358636644s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.354581216 +0000 UTC m=+145.938627843" watchObservedRunningTime="2025-10-14 06:52:09.358636644 +0000 UTC m=+145.942683271" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.392752 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" event={"ID":"4dbe6d55-1450-4f4c-8d89-bb43c7429f05","Type":"ContainerStarted","Data":"8d470a92b281f5fc3a1f2241e12effd07b78aa44a6ebd30e277b47e2603210a9"} Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396102 5018 patch_prober.go:28] interesting pod/console-operator-58897d9998-vc4k4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396143 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" podUID="d222c5fe-b248-425d-a810-83e03907f6a7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396457 5018 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6vg2h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396526 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396559 5018 patch_prober.go:28] interesting pod/downloads-7954f5f757-zvw8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.396580 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zvw8c" podUID="54e62cc6-54c6-4df2-b349-39bf04d702bc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.398556 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-jsmbp" podStartSLOduration=124.398542891 podStartE2EDuration="2m4.398542891s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.396067629 +0000 UTC m=+145.980114256" watchObservedRunningTime="2025-10-14 06:52:09.398542891 +0000 UTC m=+145.982589518" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.405911 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4wlkr" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.438338 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" podStartSLOduration=124.438309104 podStartE2EDuration="2m4.438309104s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.436187723 +0000 UTC m=+146.020234350" watchObservedRunningTime="2025-10-14 06:52:09.438309104 +0000 UTC m=+146.022355721" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.456694 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.457535 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:09.957521301 +0000 UTC m=+146.541567928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.478898 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" podStartSLOduration=124.47888096 podStartE2EDuration="2m4.47888096s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.478240982 +0000 UTC m=+146.062287619" watchObservedRunningTime="2025-10-14 06:52:09.47888096 +0000 UTC m=+146.062927587" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.519004 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dh52n" podStartSLOduration=124.518986443 podStartE2EDuration="2m4.518986443s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.51817033 +0000 UTC m=+146.102216957" watchObservedRunningTime="2025-10-14 06:52:09.518986443 +0000 UTC m=+146.103033070" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.543596 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:09 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:09 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:09 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.543696 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.559344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.561915 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.061900688 +0000 UTC m=+146.645947315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.607703 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jzjvj" podStartSLOduration=124.607688765 podStartE2EDuration="2m4.607688765s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.60543452 +0000 UTC m=+146.189481147" watchObservedRunningTime="2025-10-14 06:52:09.607688765 +0000 UTC m=+146.191735392" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.662237 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.662665 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.162642238 +0000 UTC m=+146.746688865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.662890 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.663313 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.163305438 +0000 UTC m=+146.747352065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.688045 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vmv82" podStartSLOduration=124.688020334 podStartE2EDuration="2m4.688020334s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.682903596 +0000 UTC m=+146.266950253" watchObservedRunningTime="2025-10-14 06:52:09.688020334 +0000 UTC m=+146.272066961" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.697699 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r744g" podStartSLOduration=124.697680674 podStartE2EDuration="2m4.697680674s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:09.64096509 +0000 UTC m=+146.225011717" watchObservedRunningTime="2025-10-14 06:52:09.697680674 +0000 UTC m=+146.281727301" Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.764647 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.765002 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.264987746 +0000 UTC m=+146.849034373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.865910 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.866229 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.366218261 +0000 UTC m=+146.950264888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:09 crc kubenswrapper[5018]: I1014 06:52:09.967032 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:09 crc kubenswrapper[5018]: E1014 06:52:09.967298 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.467283701 +0000 UTC m=+147.051330328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.068673 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.069043 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.569023431 +0000 UTC m=+147.153070058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.170035 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.170545 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.670529384 +0000 UTC m=+147.254576011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.271508 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.271828 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.77181679 +0000 UTC m=+147.355863417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.373185 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.373451 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.873438016 +0000 UTC m=+147.457484643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.399576 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" event={"ID":"65aea222-41ff-4806-ab92-ee7d19167a82","Type":"ContainerStarted","Data":"c6167839a4492006934609b4465a41095f640a4cd3f01dac43218793650ee572"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.402147 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" event={"ID":"c6385ef6-72b4-49f4-b61b-546be9da202c","Type":"ContainerStarted","Data":"96e1dddb6d0a4c2fa4a07f53a2c032a71b564bf36b79546525efde17931a9c33"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.404266 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" event={"ID":"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4","Type":"ContainerStarted","Data":"c6ef65970706025a9fc70d5dfdb48a67a1fb29893567be2ca7c19bd2128a23d1"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.404291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" event={"ID":"c9a1d4cc-b532-44e2-ad14-d57f87be1bd4","Type":"ContainerStarted","Data":"16e2a88044f60b21d5fce596ea2323738c08c9e6cf70b99413fd53212a25b674"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.405936 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x2z8l" event={"ID":"1357a110-9b7b-4722-9985-31cb6140b37d","Type":"ContainerStarted","Data":"2fccc995c8c695618b4601bac9503cad7523ee66fd545eeeb033ec2eeb528138"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.406317 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.408463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" event={"ID":"f81cde88-94b2-4d71-aaec-765256fe89d1","Type":"ContainerStarted","Data":"26ecf03ed926a981372a3a83b070164877414cc0622f2b8bf46a782a64b0d355"} Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.412807 5018 patch_prober.go:28] interesting pod/downloads-7954f5f757-zvw8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.412876 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zvw8c" podUID="54e62cc6-54c6-4df2-b349-39bf04d702bc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.416092 5018 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mmmgm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.416257 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.424790 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.426012 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f9c54" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.428586 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jf2w" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.430104 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vc4k4" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.473717 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8mzkh" podStartSLOduration=125.473699683 podStartE2EDuration="2m5.473699683s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:10.47081643 +0000 UTC m=+147.054863057" watchObservedRunningTime="2025-10-14 06:52:10.473699683 +0000 UTC m=+147.057746310" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.475579 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.477870 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:10.977860374 +0000 UTC m=+147.561907001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.537076 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:10 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:10 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:10 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.537121 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.576500 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.576639 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.076608417 +0000 UTC m=+147.660655044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.576710 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.578019 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.077999688 +0000 UTC m=+147.662046385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.681918 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzlrt" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.684353 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.684573 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.184547527 +0000 UTC m=+147.768594154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.684657 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.685037 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.185023701 +0000 UTC m=+147.769070328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.712736 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x2z8l" podStartSLOduration=7.712715684 podStartE2EDuration="7.712715684s" podCreationTimestamp="2025-10-14 06:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:10.710765607 +0000 UTC m=+147.294812234" watchObservedRunningTime="2025-10-14 06:52:10.712715684 +0000 UTC m=+147.296762321" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.785868 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.786269 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.286253826 +0000 UTC m=+147.870300453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.858001 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" podStartSLOduration=126.857986176 podStartE2EDuration="2m6.857986176s" podCreationTimestamp="2025-10-14 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:10.847663466 +0000 UTC m=+147.431710093" watchObservedRunningTime="2025-10-14 06:52:10.857986176 +0000 UTC m=+147.442032803" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.859064 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" podStartSLOduration=125.859055967 podStartE2EDuration="2m5.859055967s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:10.803217568 +0000 UTC m=+147.387264195" watchObservedRunningTime="2025-10-14 06:52:10.859055967 +0000 UTC m=+147.443102594" Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.887890 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.888171 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.38816005 +0000 UTC m=+147.972206677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:10 crc kubenswrapper[5018]: I1014 06:52:10.989174 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:10 crc kubenswrapper[5018]: E1014 06:52:10.989491 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.489475678 +0000 UTC m=+148.073522305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.090676 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.091084 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.591067174 +0000 UTC m=+148.175113801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.193082 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.193258 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.693232326 +0000 UTC m=+148.277278953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.193662 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.193969 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.693961417 +0000 UTC m=+148.278008044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.294557 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.294919 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.794904984 +0000 UTC m=+148.378951611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.396351 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.396660 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.896647404 +0000 UTC m=+148.480694031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.414998 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" event={"ID":"f81cde88-94b2-4d71-aaec-765256fe89d1","Type":"ContainerStarted","Data":"13e181f3b6fc381bb10965768f4a8d3007cfc21069e3e125be8f998e63f5267d"} Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.497676 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.498066 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.498209 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.498747 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:11.99861144 +0000 UTC m=+148.582658067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.500841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.514053 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.537716 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:11 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:11 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:11 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.537780 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.554243 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.599654 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.599958 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.599983 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.600932 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.100911886 +0000 UTC m=+148.684958613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.612041 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.615235 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.704994 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.705357 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.205341794 +0000 UTC m=+148.789388421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.808063 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.808581 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.308570987 +0000 UTC m=+148.892617614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.836935 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.852925 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 06:52:11 crc kubenswrapper[5018]: I1014 06:52:11.908973 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:11 crc kubenswrapper[5018]: E1014 06:52:11.909355 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.409335529 +0000 UTC m=+148.993382156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.012078 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.012350 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.512337615 +0000 UTC m=+149.096384242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.127025 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.127269 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.627241157 +0000 UTC m=+149.211287784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.127570 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.127893 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.627882035 +0000 UTC m=+149.211928662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.229092 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.229420 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.729404619 +0000 UTC m=+149.313451246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.272967 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.292406 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.294665 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.300603 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.321381 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.322115 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.330005 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.330333 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.830320475 +0000 UTC m=+149.414367112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.334824 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.336889 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.347000 5018 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.387584 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.430568 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431117 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431160 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs8rr\" (UniqueName: \"kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.431205 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.931178 +0000 UTC m=+149.515224637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431233 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431264 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.431306 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.431646 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:12.931611012 +0000 UTC m=+149.515657639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.460916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" event={"ID":"f81cde88-94b2-4d71-aaec-765256fe89d1","Type":"ContainerStarted","Data":"637be2daf649b1edfa0487fb699883bdbf0ec47704d578f44419c8c51414af08"} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.461275 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" event={"ID":"f81cde88-94b2-4d71-aaec-765256fe89d1","Type":"ContainerStarted","Data":"eabec2d8283a6a00b742f980c158ae5b2f508623c054abd5d3c956eeff087b58"} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.469797 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2a2d844707f24ee3ba4e42c45bdb7e538e17f5b0579ca520b6922bf6773c9951"} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.475031 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.477565 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.482858 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.492372 5018 generic.go:334] "Generic (PLEG): container finished" podID="b7a8f667-d3bf-43eb-b046-291b70da5c13" containerID="24c829c769ae1e7e691b4788880cac2bc276230b80d85df18f092f5b48c7ba7d" exitCode=0 Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.492439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" event={"ID":"b7a8f667-d3bf-43eb-b046-291b70da5c13","Type":"ContainerDied","Data":"24c829c769ae1e7e691b4788880cac2bc276230b80d85df18f092f5b48c7ba7d"} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.494473 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a98e98ff1b3e1c29e83e897c1c4c3ce17bd498933a645e0f820afafcc05e3315"} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.502471 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533026 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533111 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4j2t6" podStartSLOduration=9.533093315 podStartE2EDuration="9.533093315s" podCreationTimestamp="2025-10-14 06:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:12.532340433 +0000 UTC m=+149.116387060" watchObservedRunningTime="2025-10-14 06:52:12.533093315 +0000 UTC m=+149.117139942" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533318 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533370 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533401 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs8rr\" (UniqueName: \"kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533442 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533471 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.533930 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.534016 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:13.033996671 +0000 UTC m=+149.618043298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.534286 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.534669 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.537866 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:12 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:12 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:12 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.537939 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.559446 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.574084 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs8rr\" (UniqueName: \"kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr\") pod \"community-operators-9lcsh\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.629012 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.635193 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.635289 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.635360 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.635461 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmgxp\" (UniqueName: \"kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.638049 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 06:52:13.138032148 +0000 UTC m=+149.722078775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hkvfr" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.665674 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.666613 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.685522 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.707910 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.736212 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.736663 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.736711 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmgxp\" (UniqueName: \"kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.736750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.737866 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.738138 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: E1014 06:52:12.738196 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 06:52:13.238183411 +0000 UTC m=+149.822230038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.760512 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmgxp\" (UniqueName: \"kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp\") pod \"certified-operators-g9d9j\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.786114 5018 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-14T06:52:12.347117032Z","Handler":null,"Name":""} Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.819838 5018 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.819868 5018 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.840275 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.840314 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq27v\" (UniqueName: \"kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.840335 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.840359 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.840888 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.857462 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.857500 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.867177 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.868121 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.884514 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.951369 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.951408 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq27v\" (UniqueName: \"kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.951433 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.951924 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.952218 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:12 crc kubenswrapper[5018]: I1014 06:52:12.995438 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hkvfr\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.004441 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq27v\" (UniqueName: \"kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v\") pod \"community-operators-4hgcf\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.028233 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:52:13 crc kubenswrapper[5018]: W1014 06:52:13.039562 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5ace222_4f8c_4058_8497_cb8f12b491fb.slice/crio-997a71f2facbe6177ea4554c491ffaf1caa6c5d15db2c9cd591d0465b337feb5 WatchSource:0}: Error finding container 997a71f2facbe6177ea4554c491ffaf1caa6c5d15db2c9cd591d0465b337feb5: Status 404 returned error can't find the container with id 997a71f2facbe6177ea4554c491ffaf1caa6c5d15db2c9cd591d0465b337feb5 Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.052354 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.052598 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.052643 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.052684 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xgxx\" (UniqueName: \"kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.076075 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.116919 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.153948 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.154020 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xgxx\" (UniqueName: \"kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.154079 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.154546 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.154781 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.199595 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xgxx\" (UniqueName: \"kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx\") pod \"certified-operators-bxmrz\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.205377 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.227878 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.282076 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.357244 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.516359 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e1fa37c9c4f632479b5de85226e4274899bff4538731eb24682cda2b0631c3de"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.526226 5018 generic.go:334] "Generic (PLEG): container finished" podID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerID="4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6" exitCode=0 Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.536404 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerDied","Data":"4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.536464 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerStarted","Data":"997a71f2facbe6177ea4554c491ffaf1caa6c5d15db2c9cd591d0465b337feb5"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.541455 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:13 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:13 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:13 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.541512 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.581037 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.618246 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a6a96db8c9da5a7004768c305428b90b81d65f9d501c44d3405c7272371a5789"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.618534 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20208e71fcecb779d09678109b64fe37bc723d58fe8b65af5a60eae27d6959ab"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.630156 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e785eb78-c381-4f8f-877b-8aada02836aa","Type":"ContainerStarted","Data":"a72f502fc79c492fc3186a456ed5c545120a2bec0e1e651336f086ee75ab1a13"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.666291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerStarted","Data":"a9f4488145961e600a13a637906827d6c0da1a0fc293223c5f57d8d2a208cc40"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.668603 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7c9046fe6b057429c49f1a35d4cc07d0b185e868a03d4609449479648a6854e4"} Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.668879 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.690141 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:52:13 crc kubenswrapper[5018]: W1014 06:52:13.720427 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c6accb3_b364_4b05_97b7_ac8cdcb8db31.slice/crio-7a0dfbfa84b8b327a8c248f2d3a60ae2dc57fc8faa7daaba23ec8f9f315d89f5 WatchSource:0}: Error finding container 7a0dfbfa84b8b327a8c248f2d3a60ae2dc57fc8faa7daaba23ec8f9f315d89f5: Status 404 returned error can't find the container with id 7a0dfbfa84b8b327a8c248f2d3a60ae2dc57fc8faa7daaba23ec8f9f315d89f5 Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.722608 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:13 crc kubenswrapper[5018]: W1014 06:52:13.735895 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod711f5257_2feb_4174_a5f7_2421db35b491.slice/crio-daaee04670c3a1da14e0a06acf6a84fc961ec92fea99b652aa6c922a7b03bb31 WatchSource:0}: Error finding container daaee04670c3a1da14e0a06acf6a84fc961ec92fea99b652aa6c922a7b03bb31: Status 404 returned error can't find the container with id daaee04670c3a1da14e0a06acf6a84fc961ec92fea99b652aa6c922a7b03bb31 Oct 14 06:52:13 crc kubenswrapper[5018]: I1014 06:52:13.904929 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.000334 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:14 crc kubenswrapper[5018]: W1014 06:52:14.007771 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbc5ffdf_e4df_48b4_a227_5cfbc470eb95.slice/crio-c446d27186c3afd27f85529ad9516e6db36251de8f2af75f14e7d7594a196cb1 WatchSource:0}: Error finding container c446d27186c3afd27f85529ad9516e6db36251de8f2af75f14e7d7594a196cb1: Status 404 returned error can't find the container with id c446d27186c3afd27f85529ad9516e6db36251de8f2af75f14e7d7594a196cb1 Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.080511 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume\") pod \"b7a8f667-d3bf-43eb-b046-291b70da5c13\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.080895 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt9kg\" (UniqueName: \"kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg\") pod \"b7a8f667-d3bf-43eb-b046-291b70da5c13\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.080933 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume\") pod \"b7a8f667-d3bf-43eb-b046-291b70da5c13\" (UID: \"b7a8f667-d3bf-43eb-b046-291b70da5c13\") " Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.081681 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume" (OuterVolumeSpecName: "config-volume") pod "b7a8f667-d3bf-43eb-b046-291b70da5c13" (UID: "b7a8f667-d3bf-43eb-b046-291b70da5c13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.086802 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg" (OuterVolumeSpecName: "kube-api-access-qt9kg") pod "b7a8f667-d3bf-43eb-b046-291b70da5c13" (UID: "b7a8f667-d3bf-43eb-b046-291b70da5c13"). InnerVolumeSpecName "kube-api-access-qt9kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.088500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b7a8f667-d3bf-43eb-b046-291b70da5c13" (UID: "b7a8f667-d3bf-43eb-b046-291b70da5c13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.182802 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7a8f667-d3bf-43eb-b046-291b70da5c13-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.182838 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt9kg\" (UniqueName: \"kubernetes.io/projected/b7a8f667-d3bf-43eb-b046-291b70da5c13-kube-api-access-qt9kg\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.182852 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7a8f667-d3bf-43eb-b046-291b70da5c13-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.454476 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:52:14 crc kubenswrapper[5018]: E1014 06:52:14.454807 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a8f667-d3bf-43eb-b046-291b70da5c13" containerName="collect-profiles" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.454831 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a8f667-d3bf-43eb-b046-291b70da5c13" containerName="collect-profiles" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.454953 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a8f667-d3bf-43eb-b046-291b70da5c13" containerName="collect-profiles" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.455807 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.458159 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.467186 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.536782 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:14 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:14 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:14 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.536829 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.588046 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpwfp\" (UniqueName: \"kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.588092 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.588174 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.611829 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.676434 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerID="7d25e11382c895984f7befaabc8755b28056beff442de83e15f1faaf50bed231" exitCode=0 Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.676493 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerDied","Data":"7d25e11382c895984f7befaabc8755b28056beff442de83e15f1faaf50bed231"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.676562 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerStarted","Data":"c446d27186c3afd27f85529ad9516e6db36251de8f2af75f14e7d7594a196cb1"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.678133 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" event={"ID":"b7a8f667-d3bf-43eb-b046-291b70da5c13","Type":"ContainerDied","Data":"aee0f9ad67fd7bc4b2ce9d81ecff8b0e9d08769f48992a68c240e56e83ed9e02"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.678172 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aee0f9ad67fd7bc4b2ce9d81ecff8b0e9d08769f48992a68c240e56e83ed9e02" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.678214 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.685118 5018 generic.go:334] "Generic (PLEG): container finished" podID="e785eb78-c381-4f8f-877b-8aada02836aa" containerID="9273e0b17537ffd5fd9b6285702df9452974044db508c16222bf0da494c8a90d" exitCode=0 Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.685457 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e785eb78-c381-4f8f-877b-8aada02836aa","Type":"ContainerDied","Data":"9273e0b17537ffd5fd9b6285702df9452974044db508c16222bf0da494c8a90d"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.687349 5018 generic.go:334] "Generic (PLEG): container finished" podID="711f5257-2feb-4174-a5f7-2421db35b491" containerID="353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc" exitCode=0 Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.687391 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerDied","Data":"353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.687432 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerStarted","Data":"daaee04670c3a1da14e0a06acf6a84fc961ec92fea99b652aa6c922a7b03bb31"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.688916 5018 generic.go:334] "Generic (PLEG): container finished" podID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerID="8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe" exitCode=0 Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.688976 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerDied","Data":"8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.689065 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpwfp\" (UniqueName: \"kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.689107 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.689205 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.690108 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.690184 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.694495 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" event={"ID":"7c6accb3-b364-4b05-97b7-ac8cdcb8db31","Type":"ContainerStarted","Data":"00a6ac6361dccbf9602f5d7c710ae75d987a35d1a6f3e47982125b0cc1f6305e"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.694555 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" event={"ID":"7c6accb3-b364-4b05-97b7-ac8cdcb8db31","Type":"ContainerStarted","Data":"7a0dfbfa84b8b327a8c248f2d3a60ae2dc57fc8faa7daaba23ec8f9f315d89f5"} Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.694739 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.722507 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpwfp\" (UniqueName: \"kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp\") pod \"redhat-marketplace-qmnmj\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.804782 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" podStartSLOduration=129.80476467 podStartE2EDuration="2m9.80476467s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:14.786899732 +0000 UTC m=+151.370946379" watchObservedRunningTime="2025-10-14 06:52:14.80476467 +0000 UTC m=+151.388811297" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.814980 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.851694 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.852860 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.862451 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.992587 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.992896 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjj79\" (UniqueName: \"kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:14 crc kubenswrapper[5018]: I1014 06:52:14.992981 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.024400 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:52:15 crc kubenswrapper[5018]: W1014 06:52:15.031258 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a7c05d_8be0_4bc3_8f60_06d8abdfa9a3.slice/crio-606cb3b384febea3e0f5562cb8eb42f31dbb923c38d6a4ed6ee66981e89f7fa6 WatchSource:0}: Error finding container 606cb3b384febea3e0f5562cb8eb42f31dbb923c38d6a4ed6ee66981e89f7fa6: Status 404 returned error can't find the container with id 606cb3b384febea3e0f5562cb8eb42f31dbb923c38d6a4ed6ee66981e89f7fa6 Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.094562 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.094609 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjj79\" (UniqueName: \"kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.094705 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.095117 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.095320 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.128783 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjj79\" (UniqueName: \"kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79\") pod \"redhat-marketplace-k2hmk\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.178898 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.311451 5018 patch_prober.go:28] interesting pod/downloads-7954f5f757-zvw8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.311779 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zvw8c" podUID="54e62cc6-54c6-4df2-b349-39bf04d702bc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.311516 5018 patch_prober.go:28] interesting pod/downloads-7954f5f757-zvw8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.311974 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zvw8c" podUID="54e62cc6-54c6-4df2-b349-39bf04d702bc" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.372031 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.372077 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.380894 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.380949 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.381339 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.387240 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.455787 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.456707 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.465633 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.470535 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.533858 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.536899 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:15 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:15 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:15 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.536950 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.604234 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbmvl\" (UniqueName: \"kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.604274 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.604356 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.660338 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:15 crc kubenswrapper[5018]: W1014 06:52:15.667868 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02ec73f5_e0ea_4433_95cc_50e17c0329b3.slice/crio-978c1b2391b372c154ca05e7b30701042ee9859192ae32a2095bc15f4a112935 WatchSource:0}: Error finding container 978c1b2391b372c154ca05e7b30701042ee9859192ae32a2095bc15f4a112935: Status 404 returned error can't find the container with id 978c1b2391b372c154ca05e7b30701042ee9859192ae32a2095bc15f4a112935 Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.705019 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbmvl\" (UniqueName: \"kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.705062 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.705189 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.706009 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.706075 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.708890 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerStarted","Data":"978c1b2391b372c154ca05e7b30701042ee9859192ae32a2095bc15f4a112935"} Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.744016 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbmvl\" (UniqueName: \"kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl\") pod \"redhat-operators-sf5hl\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.754156 5018 generic.go:334] "Generic (PLEG): container finished" podID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerID="d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc" exitCode=0 Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.754964 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerDied","Data":"d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc"} Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.754998 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerStarted","Data":"606cb3b384febea3e0f5562cb8eb42f31dbb923c38d6a4ed6ee66981e89f7fa6"} Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.761543 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-7dthh" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.771798 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vf8lt" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.778890 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.826499 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.826775 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.846711 5018 patch_prober.go:28] interesting pod/console-f9d7485db-sc99m container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.846931 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sc99m" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.885281 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.886544 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.903693 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.906244 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.911708 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.916717 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.916840 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 14 06:52:15 crc kubenswrapper[5018]: I1014 06:52:15.927537 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.010159 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5khgs\" (UniqueName: \"kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.010203 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.010246 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.010265 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.010310 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.025893 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:52:16 crc kubenswrapper[5018]: E1014 06:52:16.083021 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02ec73f5_e0ea_4433_95cc_50e17c0329b3.slice/crio-9007267a192446e5d203075aa3ce884c814764b23d813a905b8c289f0e554292.scope\": RecentStats: unable to find data in memory cache]" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.111977 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5khgs\" (UniqueName: \"kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112027 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112079 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112098 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112148 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112587 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.112994 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.114313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.135291 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5khgs\" (UniqueName: \"kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs\") pod \"redhat-operators-jlf9g\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.137988 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.241795 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.251879 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.272347 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.277254 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.317450 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir\") pod \"e785eb78-c381-4f8f-877b-8aada02836aa\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.317551 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e785eb78-c381-4f8f-877b-8aada02836aa" (UID: "e785eb78-c381-4f8f-877b-8aada02836aa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.317600 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access\") pod \"e785eb78-c381-4f8f-877b-8aada02836aa\" (UID: \"e785eb78-c381-4f8f-877b-8aada02836aa\") " Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.317966 5018 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e785eb78-c381-4f8f-877b-8aada02836aa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.321689 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e785eb78-c381-4f8f-877b-8aada02836aa" (UID: "e785eb78-c381-4f8f-877b-8aada02836aa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.418702 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e785eb78-c381-4f8f-877b-8aada02836aa-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.547217 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:16 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:16 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:16 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.547507 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.572496 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.603893 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.794417 5018 generic.go:334] "Generic (PLEG): container finished" podID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerID="9007267a192446e5d203075aa3ce884c814764b23d813a905b8c289f0e554292" exitCode=0 Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.794501 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerDied","Data":"9007267a192446e5d203075aa3ce884c814764b23d813a905b8c289f0e554292"} Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.802254 5018 generic.go:334] "Generic (PLEG): container finished" podID="9294a087-ccc1-4955-b730-fd5f3669c190" containerID="0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663" exitCode=0 Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.802492 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerDied","Data":"0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663"} Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.802543 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerStarted","Data":"79bb01d2ebd71db655a1647a06c16e0b5e35d0fd8a28bb7769d9f6b6720839be"} Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.807127 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.807141 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e785eb78-c381-4f8f-877b-8aada02836aa","Type":"ContainerDied","Data":"a72f502fc79c492fc3186a456ed5c545120a2bec0e1e651336f086ee75ab1a13"} Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.807175 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72f502fc79c492fc3186a456ed5c545120a2bec0e1e651336f086ee75ab1a13" Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.809097 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6","Type":"ContainerStarted","Data":"791775b5aae40c8a0dc597b1ec9a285d85db0f431fa6d6b9dc26513d5bccd606"} Oct 14 06:52:16 crc kubenswrapper[5018]: I1014 06:52:16.814217 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerStarted","Data":"a1ff25c99003230d4b1d2b115e1b8ba105843f4cc76c22ea46fb3c1154679d80"} Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.540388 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:17 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:17 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:17 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.540744 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.828747 5018 generic.go:334] "Generic (PLEG): container finished" podID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerID="8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9" exitCode=0 Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.829034 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerDied","Data":"8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9"} Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.832559 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6","Type":"ContainerStarted","Data":"71d54305faf8ec75f26294005ea96279328fd3588ce982eece1074d12fa8ff98"} Oct 14 06:52:17 crc kubenswrapper[5018]: I1014 06:52:17.860591 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.8605611 podStartE2EDuration="2.8605611s" podCreationTimestamp="2025-10-14 06:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:17.859768197 +0000 UTC m=+154.443814824" watchObservedRunningTime="2025-10-14 06:52:17.8605611 +0000 UTC m=+154.444607727" Oct 14 06:52:18 crc kubenswrapper[5018]: I1014 06:52:18.085358 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x2z8l" Oct 14 06:52:18 crc kubenswrapper[5018]: I1014 06:52:18.536485 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:18 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:18 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:18 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:18 crc kubenswrapper[5018]: I1014 06:52:18.536809 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:18 crc kubenswrapper[5018]: I1014 06:52:18.843333 5018 generic.go:334] "Generic (PLEG): container finished" podID="6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" containerID="71d54305faf8ec75f26294005ea96279328fd3588ce982eece1074d12fa8ff98" exitCode=0 Oct 14 06:52:18 crc kubenswrapper[5018]: I1014 06:52:18.843382 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6","Type":"ContainerDied","Data":"71d54305faf8ec75f26294005ea96279328fd3588ce982eece1074d12fa8ff98"} Oct 14 06:52:19 crc kubenswrapper[5018]: I1014 06:52:19.535805 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:19 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:19 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:19 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:19 crc kubenswrapper[5018]: I1014 06:52:19.535883 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:20 crc kubenswrapper[5018]: I1014 06:52:20.541161 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:20 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:20 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:20 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:20 crc kubenswrapper[5018]: I1014 06:52:20.541738 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:21 crc kubenswrapper[5018]: I1014 06:52:21.536457 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:21 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:21 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:21 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:21 crc kubenswrapper[5018]: I1014 06:52:21.536533 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:22 crc kubenswrapper[5018]: I1014 06:52:22.535473 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:22 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:22 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:22 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:22 crc kubenswrapper[5018]: I1014 06:52:22.535758 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:23 crc kubenswrapper[5018]: I1014 06:52:23.535141 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:23 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:23 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:23 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:23 crc kubenswrapper[5018]: I1014 06:52:23.535184 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.396055 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.438682 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access\") pod \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.438773 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir\") pod \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\" (UID: \"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6\") " Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.438915 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" (UID: "6c64c9b4-60fc-4409-b38a-832c3e5fe5b6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.439179 5018 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.445967 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" (UID: "6c64c9b4-60fc-4409-b38a-832c3e5fe5b6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.537580 5018 patch_prober.go:28] interesting pod/router-default-5444994796-5hjlc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 06:52:24 crc kubenswrapper[5018]: [-]has-synced failed: reason withheld Oct 14 06:52:24 crc kubenswrapper[5018]: [+]process-running ok Oct 14 06:52:24 crc kubenswrapper[5018]: healthz check failed Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.537665 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5hjlc" podUID="6559b957-5c56-410a-86d6-b75422bf82b3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.540677 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6c64c9b4-60fc-4409-b38a-832c3e5fe5b6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.881776 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6c64c9b4-60fc-4409-b38a-832c3e5fe5b6","Type":"ContainerDied","Data":"791775b5aae40c8a0dc597b1ec9a285d85db0f431fa6d6b9dc26513d5bccd606"} Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.881812 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="791775b5aae40c8a0dc597b1ec9a285d85db0f431fa6d6b9dc26513d5bccd606" Oct 14 06:52:24 crc kubenswrapper[5018]: I1014 06:52:24.881850 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 06:52:25 crc kubenswrapper[5018]: I1014 06:52:25.317065 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-zvw8c" Oct 14 06:52:25 crc kubenswrapper[5018]: I1014 06:52:25.537880 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:25 crc kubenswrapper[5018]: I1014 06:52:25.542046 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5hjlc" Oct 14 06:52:25 crc kubenswrapper[5018]: I1014 06:52:25.824505 5018 patch_prober.go:28] interesting pod/console-f9d7485db-sc99m container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 14 06:52:25 crc kubenswrapper[5018]: I1014 06:52:25.824588 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sc99m" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 14 06:52:27 crc kubenswrapper[5018]: I1014 06:52:27.705190 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:52:27 crc kubenswrapper[5018]: I1014 06:52:27.713073 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2530b46d-5f3a-4495-af23-3ad85327c99c-metrics-certs\") pod \"network-metrics-daemon-gl9fg\" (UID: \"2530b46d-5f3a-4495-af23-3ad85327c99c\") " pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:52:27 crc kubenswrapper[5018]: I1014 06:52:27.722812 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gl9fg" Oct 14 06:52:32 crc kubenswrapper[5018]: I1014 06:52:32.464142 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:52:32 crc kubenswrapper[5018]: I1014 06:52:32.464703 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:52:33 crc kubenswrapper[5018]: I1014 06:52:33.122282 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:52:36 crc kubenswrapper[5018]: I1014 06:52:36.007611 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:36 crc kubenswrapper[5018]: I1014 06:52:36.013338 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.628072 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gl9fg"] Oct 14 06:52:41 crc kubenswrapper[5018]: W1014 06:52:41.888555 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2530b46d_5f3a_4495_af23_3ad85327c99c.slice/crio-7af392bb3228d3a3dac02bf458cb795820f9d4fcfd322ed4901e628762e821e7 WatchSource:0}: Error finding container 7af392bb3228d3a3dac02bf458cb795820f9d4fcfd322ed4901e628762e821e7: Status 404 returned error can't find the container with id 7af392bb3228d3a3dac02bf458cb795820f9d4fcfd322ed4901e628762e821e7 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.973729 5018 generic.go:334] "Generic (PLEG): container finished" podID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerID="a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5" exitCode=0 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.973794 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerDied","Data":"a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5"} Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.976717 5018 generic.go:334] "Generic (PLEG): container finished" podID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerID="764b48ab3533c7cf786dc1b66886fcff1cbd51ff184ce634da25a7a7c7729551" exitCode=0 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.976762 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerDied","Data":"764b48ab3533c7cf786dc1b66886fcff1cbd51ff184ce634da25a7a7c7729551"} Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.981301 5018 generic.go:334] "Generic (PLEG): container finished" podID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerID="bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c" exitCode=0 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.981363 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerDied","Data":"bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c"} Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.986612 5018 generic.go:334] "Generic (PLEG): container finished" podID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerID="69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b" exitCode=0 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.986669 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerDied","Data":"69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b"} Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.993690 5018 generic.go:334] "Generic (PLEG): container finished" podID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerID="bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241" exitCode=0 Oct 14 06:52:41 crc kubenswrapper[5018]: I1014 06:52:41.993734 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerDied","Data":"bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241"} Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.000223 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" event={"ID":"2530b46d-5f3a-4495-af23-3ad85327c99c","Type":"ContainerStarted","Data":"7af392bb3228d3a3dac02bf458cb795820f9d4fcfd322ed4901e628762e821e7"} Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.002167 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerID="64e399eb94a7a40bd7c17871798e4b4992033e2a7ec8aff15fc3572a090c025f" exitCode=0 Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.002288 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerDied","Data":"64e399eb94a7a40bd7c17871798e4b4992033e2a7ec8aff15fc3572a090c025f"} Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.008517 5018 generic.go:334] "Generic (PLEG): container finished" podID="711f5257-2feb-4174-a5f7-2421db35b491" containerID="d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546" exitCode=0 Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.008587 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerDied","Data":"d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546"} Oct 14 06:52:42 crc kubenswrapper[5018]: I1014 06:52:42.019193 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerStarted","Data":"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.031693 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerStarted","Data":"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.036427 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerStarted","Data":"f1eb801bbdd3787cd5f32ef358310df6374a94b435099ca590f965152955a606"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.040273 5018 generic.go:334] "Generic (PLEG): container finished" podID="9294a087-ccc1-4955-b730-fd5f3669c190" containerID="ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31" exitCode=0 Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.040370 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerDied","Data":"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.043510 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerStarted","Data":"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.048502 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" event={"ID":"2530b46d-5f3a-4495-af23-3ad85327c99c","Type":"ContainerStarted","Data":"9abe4121f6d901787afb067494c0f57c26b17de392c4514cc6c7e5a8db224078"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.048554 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gl9fg" event={"ID":"2530b46d-5f3a-4495-af23-3ad85327c99c","Type":"ContainerStarted","Data":"024b43298dabb984bbc4af039c719c1ae824dd06c201b2b12416d5c0e9a7c78b"} Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.061731 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jlf9g" podStartSLOduration=3.40863137 podStartE2EDuration="28.061691273s" podCreationTimestamp="2025-10-14 06:52:15 +0000 UTC" firstStartedPulling="2025-10-14 06:52:17.832030562 +0000 UTC m=+154.416077189" lastFinishedPulling="2025-10-14 06:52:42.485090455 +0000 UTC m=+179.069137092" observedRunningTime="2025-10-14 06:52:43.060707664 +0000 UTC m=+179.644754291" watchObservedRunningTime="2025-10-14 06:52:43.061691273 +0000 UTC m=+179.645737900" Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.081055 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gl9fg" podStartSLOduration=158.081038794 podStartE2EDuration="2m38.081038794s" podCreationTimestamp="2025-10-14 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:52:43.078572712 +0000 UTC m=+179.662619349" watchObservedRunningTime="2025-10-14 06:52:43.081038794 +0000 UTC m=+179.665085421" Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.136709 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9lcsh" podStartSLOduration=2.211444188 podStartE2EDuration="31.136683237s" podCreationTimestamp="2025-10-14 06:52:12 +0000 UTC" firstStartedPulling="2025-10-14 06:52:13.580603557 +0000 UTC m=+150.164650194" lastFinishedPulling="2025-10-14 06:52:42.505842616 +0000 UTC m=+179.089889243" observedRunningTime="2025-10-14 06:52:43.104743131 +0000 UTC m=+179.688789758" watchObservedRunningTime="2025-10-14 06:52:43.136683237 +0000 UTC m=+179.720729884" Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.161737 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bxmrz" podStartSLOduration=3.323262463 podStartE2EDuration="31.161714253s" podCreationTimestamp="2025-10-14 06:52:12 +0000 UTC" firstStartedPulling="2025-10-14 06:52:14.678469258 +0000 UTC m=+151.262515885" lastFinishedPulling="2025-10-14 06:52:42.516921048 +0000 UTC m=+179.100967675" observedRunningTime="2025-10-14 06:52:43.15886479 +0000 UTC m=+179.742911417" watchObservedRunningTime="2025-10-14 06:52:43.161714253 +0000 UTC m=+179.745760880" Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.229447 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:43 crc kubenswrapper[5018]: I1014 06:52:43.229509 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.068793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerStarted","Data":"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907"} Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.070949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerStarted","Data":"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326"} Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.073264 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerStarted","Data":"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4"} Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.078271 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerStarted","Data":"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf"} Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.081253 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerStarted","Data":"3400cf92378129c19411b1a815678e2342694367946b984c540aafef40c9ea71"} Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.116679 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4hgcf" podStartSLOduration=4.195134633 podStartE2EDuration="32.116657521s" podCreationTimestamp="2025-10-14 06:52:12 +0000 UTC" firstStartedPulling="2025-10-14 06:52:14.688666254 +0000 UTC m=+151.272712881" lastFinishedPulling="2025-10-14 06:52:42.610189122 +0000 UTC m=+179.194235769" observedRunningTime="2025-10-14 06:52:44.093543921 +0000 UTC m=+180.677590548" watchObservedRunningTime="2025-10-14 06:52:44.116657521 +0000 UTC m=+180.700704148" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.147040 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sf5hl" podStartSLOduration=2.4135307089999998 podStartE2EDuration="29.147020191s" podCreationTimestamp="2025-10-14 06:52:15 +0000 UTC" firstStartedPulling="2025-10-14 06:52:16.824229423 +0000 UTC m=+153.408276060" lastFinishedPulling="2025-10-14 06:52:43.557718915 +0000 UTC m=+180.141765542" observedRunningTime="2025-10-14 06:52:44.120231954 +0000 UTC m=+180.704278581" watchObservedRunningTime="2025-10-14 06:52:44.147020191 +0000 UTC m=+180.731066828" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.147308 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g9d9j" podStartSLOduration=3.942269711 podStartE2EDuration="32.147300189s" podCreationTimestamp="2025-10-14 06:52:12 +0000 UTC" firstStartedPulling="2025-10-14 06:52:14.689940411 +0000 UTC m=+151.273987058" lastFinishedPulling="2025-10-14 06:52:42.894970909 +0000 UTC m=+179.479017536" observedRunningTime="2025-10-14 06:52:44.135417205 +0000 UTC m=+180.719463852" watchObservedRunningTime="2025-10-14 06:52:44.147300189 +0000 UTC m=+180.731346816" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.155832 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qmnmj" podStartSLOduration=2.871821163 podStartE2EDuration="30.155818156s" podCreationTimestamp="2025-10-14 06:52:14 +0000 UTC" firstStartedPulling="2025-10-14 06:52:15.756113384 +0000 UTC m=+152.340160011" lastFinishedPulling="2025-10-14 06:52:43.040110377 +0000 UTC m=+179.624157004" observedRunningTime="2025-10-14 06:52:44.152348966 +0000 UTC m=+180.736395613" watchObservedRunningTime="2025-10-14 06:52:44.155818156 +0000 UTC m=+180.739864783" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.175690 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2hmk" podStartSLOduration=4.058569712 podStartE2EDuration="30.175669692s" podCreationTimestamp="2025-10-14 06:52:14 +0000 UTC" firstStartedPulling="2025-10-14 06:52:16.797591441 +0000 UTC m=+153.381638068" lastFinishedPulling="2025-10-14 06:52:42.914691381 +0000 UTC m=+179.498738048" observedRunningTime="2025-10-14 06:52:44.173377055 +0000 UTC m=+180.757423682" watchObservedRunningTime="2025-10-14 06:52:44.175669692 +0000 UTC m=+180.759716319" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.426892 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-bxmrz" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="registry-server" probeResult="failure" output=< Oct 14 06:52:44 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 06:52:44 crc kubenswrapper[5018]: > Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.816786 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.817029 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:44 crc kubenswrapper[5018]: I1014 06:52:44.879079 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:45 crc kubenswrapper[5018]: I1014 06:52:45.179597 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:45 crc kubenswrapper[5018]: I1014 06:52:45.179701 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:45 crc kubenswrapper[5018]: I1014 06:52:45.221310 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:45 crc kubenswrapper[5018]: I1014 06:52:45.779963 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:45 crc kubenswrapper[5018]: I1014 06:52:45.780491 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:46 crc kubenswrapper[5018]: I1014 06:52:46.272790 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:46 crc kubenswrapper[5018]: I1014 06:52:46.273852 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:46 crc kubenswrapper[5018]: I1014 06:52:46.325663 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zfjfk" Oct 14 06:52:46 crc kubenswrapper[5018]: I1014 06:52:46.821853 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sf5hl" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="registry-server" probeResult="failure" output=< Oct 14 06:52:46 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 06:52:46 crc kubenswrapper[5018]: > Oct 14 06:52:47 crc kubenswrapper[5018]: I1014 06:52:47.321497 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jlf9g" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="registry-server" probeResult="failure" output=< Oct 14 06:52:47 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 06:52:47 crc kubenswrapper[5018]: > Oct 14 06:52:51 crc kubenswrapper[5018]: I1014 06:52:51.844185 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.630054 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.630133 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.702744 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.846986 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.847064 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:52 crc kubenswrapper[5018]: I1014 06:52:52.915332 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.199948 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.205379 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.282382 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.282667 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.310598 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.329952 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:53 crc kubenswrapper[5018]: I1014 06:52:53.355274 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:54 crc kubenswrapper[5018]: I1014 06:52:54.176686 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:54 crc kubenswrapper[5018]: I1014 06:52:54.543198 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:54 crc kubenswrapper[5018]: I1014 06:52:54.872177 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:52:55 crc kubenswrapper[5018]: I1014 06:52:55.145355 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bxmrz" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="registry-server" containerID="cri-o://f1eb801bbdd3787cd5f32ef358310df6374a94b435099ca590f965152955a606" gracePeriod=2 Oct 14 06:52:55 crc kubenswrapper[5018]: I1014 06:52:55.244919 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:55 crc kubenswrapper[5018]: I1014 06:52:55.542919 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:55 crc kubenswrapper[5018]: I1014 06:52:55.833551 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:55 crc kubenswrapper[5018]: I1014 06:52:55.875598 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.152728 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerID="f1eb801bbdd3787cd5f32ef358310df6374a94b435099ca590f965152955a606" exitCode=0 Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.153775 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerDied","Data":"f1eb801bbdd3787cd5f32ef358310df6374a94b435099ca590f965152955a606"} Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.241133 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.307120 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.388515 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content\") pod \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.388640 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xgxx\" (UniqueName: \"kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx\") pod \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.388664 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities\") pod \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\" (UID: \"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95\") " Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.397812 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities" (OuterVolumeSpecName: "utilities") pod "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" (UID: "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.401862 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx" (OuterVolumeSpecName: "kube-api-access-6xgxx") pod "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" (UID: "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95"). InnerVolumeSpecName "kube-api-access-6xgxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.408094 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.439301 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" (UID: "fbc5ffdf-e4df-48b4-a227-5cfbc470eb95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.490458 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xgxx\" (UniqueName: \"kubernetes.io/projected/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-kube-api-access-6xgxx\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.490493 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:56 crc kubenswrapper[5018]: I1014 06:52:56.490505 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.159788 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxmrz" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.159890 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4hgcf" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="registry-server" containerID="cri-o://4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907" gracePeriod=2 Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.159852 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxmrz" event={"ID":"fbc5ffdf-e4df-48b4-a227-5cfbc470eb95","Type":"ContainerDied","Data":"c446d27186c3afd27f85529ad9516e6db36251de8f2af75f14e7d7594a196cb1"} Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.160672 5018 scope.go:117] "RemoveContainer" containerID="f1eb801bbdd3787cd5f32ef358310df6374a94b435099ca590f965152955a606" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.180858 5018 scope.go:117] "RemoveContainer" containerID="64e399eb94a7a40bd7c17871798e4b4992033e2a7ec8aff15fc3572a090c025f" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.184760 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.188274 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bxmrz"] Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.197302 5018 scope.go:117] "RemoveContainer" containerID="7d25e11382c895984f7befaabc8755b28056beff442de83e15f1faaf50bed231" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.558305 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.716325 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities\") pod \"711f5257-2feb-4174-a5f7-2421db35b491\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.716399 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq27v\" (UniqueName: \"kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v\") pod \"711f5257-2feb-4174-a5f7-2421db35b491\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.716460 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content\") pod \"711f5257-2feb-4174-a5f7-2421db35b491\" (UID: \"711f5257-2feb-4174-a5f7-2421db35b491\") " Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.717116 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities" (OuterVolumeSpecName: "utilities") pod "711f5257-2feb-4174-a5f7-2421db35b491" (UID: "711f5257-2feb-4174-a5f7-2421db35b491"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.724414 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v" (OuterVolumeSpecName: "kube-api-access-gq27v") pod "711f5257-2feb-4174-a5f7-2421db35b491" (UID: "711f5257-2feb-4174-a5f7-2421db35b491"). InnerVolumeSpecName "kube-api-access-gq27v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.762329 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "711f5257-2feb-4174-a5f7-2421db35b491" (UID: "711f5257-2feb-4174-a5f7-2421db35b491"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.818167 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq27v\" (UniqueName: \"kubernetes.io/projected/711f5257-2feb-4174-a5f7-2421db35b491-kube-api-access-gq27v\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.818194 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.818203 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711f5257-2feb-4174-a5f7-2421db35b491-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.937602 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:57 crc kubenswrapper[5018]: I1014 06:52:57.937878 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k2hmk" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="registry-server" containerID="cri-o://3400cf92378129c19411b1a815678e2342694367946b984c540aafef40c9ea71" gracePeriod=2 Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.166984 5018 generic.go:334] "Generic (PLEG): container finished" podID="711f5257-2feb-4174-a5f7-2421db35b491" containerID="4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907" exitCode=0 Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.167048 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hgcf" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.167053 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerDied","Data":"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907"} Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.167108 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hgcf" event={"ID":"711f5257-2feb-4174-a5f7-2421db35b491","Type":"ContainerDied","Data":"daaee04670c3a1da14e0a06acf6a84fc961ec92fea99b652aa6c922a7b03bb31"} Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.167131 5018 scope.go:117] "RemoveContainer" containerID="4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.170400 5018 generic.go:334] "Generic (PLEG): container finished" podID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerID="3400cf92378129c19411b1a815678e2342694367946b984c540aafef40c9ea71" exitCode=0 Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.170471 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerDied","Data":"3400cf92378129c19411b1a815678e2342694367946b984c540aafef40c9ea71"} Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.193296 5018 scope.go:117] "RemoveContainer" containerID="d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.197372 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.210532 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4hgcf"] Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.211224 5018 scope.go:117] "RemoveContainer" containerID="353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.229380 5018 scope.go:117] "RemoveContainer" containerID="4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907" Oct 14 06:52:58 crc kubenswrapper[5018]: E1014 06:52:58.229841 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907\": container with ID starting with 4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907 not found: ID does not exist" containerID="4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.229880 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907"} err="failed to get container status \"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907\": rpc error: code = NotFound desc = could not find container \"4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907\": container with ID starting with 4617391fd54dc77d3612980ecf4af0ccdfbf1e195800a74e94e7e02fa2d45907 not found: ID does not exist" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.229927 5018 scope.go:117] "RemoveContainer" containerID="d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546" Oct 14 06:52:58 crc kubenswrapper[5018]: E1014 06:52:58.230206 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546\": container with ID starting with d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546 not found: ID does not exist" containerID="d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.230235 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546"} err="failed to get container status \"d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546\": rpc error: code = NotFound desc = could not find container \"d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546\": container with ID starting with d23874b912f15567ed6902eb77ac91cee4dcc8c2f986898d63d41a626f68c546 not found: ID does not exist" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.230252 5018 scope.go:117] "RemoveContainer" containerID="353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc" Oct 14 06:52:58 crc kubenswrapper[5018]: E1014 06:52:58.230658 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc\": container with ID starting with 353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc not found: ID does not exist" containerID="353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.230686 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc"} err="failed to get container status \"353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc\": rpc error: code = NotFound desc = could not find container \"353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc\": container with ID starting with 353bebf1405ada7ecf9e4b984f4b1b5c91fec32f9fef2f015e26af93273f21cc not found: ID does not exist" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.274244 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.424828 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content\") pod \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.424870 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjj79\" (UniqueName: \"kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79\") pod \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.424891 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities\") pod \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\" (UID: \"02ec73f5-e0ea-4433-95cc-50e17c0329b3\") " Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.425816 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities" (OuterVolumeSpecName: "utilities") pod "02ec73f5-e0ea-4433-95cc-50e17c0329b3" (UID: "02ec73f5-e0ea-4433-95cc-50e17c0329b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.431406 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79" (OuterVolumeSpecName: "kube-api-access-cjj79") pod "02ec73f5-e0ea-4433-95cc-50e17c0329b3" (UID: "02ec73f5-e0ea-4433-95cc-50e17c0329b3"). InnerVolumeSpecName "kube-api-access-cjj79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.461072 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02ec73f5-e0ea-4433-95cc-50e17c0329b3" (UID: "02ec73f5-e0ea-4433-95cc-50e17c0329b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.526091 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.526141 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjj79\" (UniqueName: \"kubernetes.io/projected/02ec73f5-e0ea-4433-95cc-50e17c0329b3-kube-api-access-cjj79\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.526162 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02ec73f5-e0ea-4433-95cc-50e17c0329b3-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.611145 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711f5257-2feb-4174-a5f7-2421db35b491" path="/var/lib/kubelet/pods/711f5257-2feb-4174-a5f7-2421db35b491/volumes" Oct 14 06:52:58 crc kubenswrapper[5018]: I1014 06:52:58.611888 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" path="/var/lib/kubelet/pods/fbc5ffdf-e4df-48b4-a227-5cfbc470eb95/volumes" Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.179367 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2hmk" event={"ID":"02ec73f5-e0ea-4433-95cc-50e17c0329b3","Type":"ContainerDied","Data":"978c1b2391b372c154ca05e7b30701042ee9859192ae32a2095bc15f4a112935"} Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.179448 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2hmk" Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.179933 5018 scope.go:117] "RemoveContainer" containerID="3400cf92378129c19411b1a815678e2342694367946b984c540aafef40c9ea71" Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.200438 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.203283 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2hmk"] Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.210432 5018 scope.go:117] "RemoveContainer" containerID="764b48ab3533c7cf786dc1b66886fcff1cbd51ff184ce634da25a7a7c7729551" Oct 14 06:52:59 crc kubenswrapper[5018]: I1014 06:52:59.225423 5018 scope.go:117] "RemoveContainer" containerID="9007267a192446e5d203075aa3ce884c814764b23d813a905b8c289f0e554292" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.343533 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.344670 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jlf9g" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="registry-server" containerID="cri-o://7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757" gracePeriod=2 Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.614356 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" path="/var/lib/kubelet/pods/02ec73f5-e0ea-4433-95cc-50e17c0329b3/volumes" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.718750 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.863075 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content\") pod \"7f265381-35e5-4c3c-a75d-befe5443ca58\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.863127 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities\") pod \"7f265381-35e5-4c3c-a75d-befe5443ca58\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.863150 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5khgs\" (UniqueName: \"kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs\") pod \"7f265381-35e5-4c3c-a75d-befe5443ca58\" (UID: \"7f265381-35e5-4c3c-a75d-befe5443ca58\") " Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.864333 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities" (OuterVolumeSpecName: "utilities") pod "7f265381-35e5-4c3c-a75d-befe5443ca58" (UID: "7f265381-35e5-4c3c-a75d-befe5443ca58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.869658 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs" (OuterVolumeSpecName: "kube-api-access-5khgs") pod "7f265381-35e5-4c3c-a75d-befe5443ca58" (UID: "7f265381-35e5-4c3c-a75d-befe5443ca58"). InnerVolumeSpecName "kube-api-access-5khgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.964675 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5khgs\" (UniqueName: \"kubernetes.io/projected/7f265381-35e5-4c3c-a75d-befe5443ca58-kube-api-access-5khgs\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.964704 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:00 crc kubenswrapper[5018]: I1014 06:53:00.969271 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f265381-35e5-4c3c-a75d-befe5443ca58" (UID: "7f265381-35e5-4c3c-a75d-befe5443ca58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.066904 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f265381-35e5-4c3c-a75d-befe5443ca58-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.194345 5018 generic.go:334] "Generic (PLEG): container finished" podID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerID="7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757" exitCode=0 Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.194386 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerDied","Data":"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757"} Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.194415 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jlf9g" event={"ID":"7f265381-35e5-4c3c-a75d-befe5443ca58","Type":"ContainerDied","Data":"a1ff25c99003230d4b1d2b115e1b8ba105843f4cc76c22ea46fb3c1154679d80"} Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.194436 5018 scope.go:117] "RemoveContainer" containerID="7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.194455 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jlf9g" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.213682 5018 scope.go:117] "RemoveContainer" containerID="a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.229456 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.235957 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jlf9g"] Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.241969 5018 scope.go:117] "RemoveContainer" containerID="8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.258139 5018 scope.go:117] "RemoveContainer" containerID="7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757" Oct 14 06:53:01 crc kubenswrapper[5018]: E1014 06:53:01.259048 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757\": container with ID starting with 7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757 not found: ID does not exist" containerID="7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.259218 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757"} err="failed to get container status \"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757\": rpc error: code = NotFound desc = could not find container \"7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757\": container with ID starting with 7ff31af25b74f2417964d718c09f6674ac3a2ec31f4b015ca28946401b26e757 not found: ID does not exist" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.259251 5018 scope.go:117] "RemoveContainer" containerID="a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5" Oct 14 06:53:01 crc kubenswrapper[5018]: E1014 06:53:01.259661 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5\": container with ID starting with a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5 not found: ID does not exist" containerID="a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.259693 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5"} err="failed to get container status \"a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5\": rpc error: code = NotFound desc = could not find container \"a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5\": container with ID starting with a507c84dc4b928a261857190dd1457582fee59ba08608cf6feb316de9683a3f5 not found: ID does not exist" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.259716 5018 scope.go:117] "RemoveContainer" containerID="8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9" Oct 14 06:53:01 crc kubenswrapper[5018]: E1014 06:53:01.260194 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9\": container with ID starting with 8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9 not found: ID does not exist" containerID="8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9" Oct 14 06:53:01 crc kubenswrapper[5018]: I1014 06:53:01.260231 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9"} err="failed to get container status \"8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9\": rpc error: code = NotFound desc = could not find container \"8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9\": container with ID starting with 8bec0f07a55583ded251c3c38bb24c5861171995f94a8ae7ff4592ac419b33c9 not found: ID does not exist" Oct 14 06:53:02 crc kubenswrapper[5018]: I1014 06:53:02.463362 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:53:02 crc kubenswrapper[5018]: I1014 06:53:02.463727 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:53:02 crc kubenswrapper[5018]: I1014 06:53:02.611930 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" path="/var/lib/kubelet/pods/7f265381-35e5-4c3c-a75d-befe5443ca58/volumes" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.754495 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.756024 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g9d9j" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="registry-server" containerID="cri-o://27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4" gracePeriod=30 Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.763319 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.763813 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9lcsh" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="registry-server" containerID="cri-o://92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68" gracePeriod=30 Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.775008 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.775229 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" containerID="cri-o://8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199" gracePeriod=30 Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.791648 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.791958 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qmnmj" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="registry-server" containerID="cri-o://503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf" gracePeriod=30 Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797344 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bh2kb"] Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797661 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797677 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797693 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797702 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797711 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797719 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797734 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797740 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797750 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797757 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797767 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797774 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797804 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797812 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797820 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797827 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797836 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797842 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="extract-utilities" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797853 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e785eb78-c381-4f8f-877b-8aada02836aa" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797860 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e785eb78-c381-4f8f-877b-8aada02836aa" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797870 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797878 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797888 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797895 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797905 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797913 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: E1014 06:53:30.797922 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.797930 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="extract-content" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798049 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f265381-35e5-4c3c-a75d-befe5443ca58" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798062 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="711f5257-2feb-4174-a5f7-2421db35b491" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798070 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c64c9b4-60fc-4409-b38a-832c3e5fe5b6" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798081 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e785eb78-c381-4f8f-877b-8aada02836aa" containerName="pruner" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798091 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ec73f5-e0ea-4433-95cc-50e17c0329b3" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798100 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc5ffdf-e4df-48b4-a227-5cfbc470eb95" containerName="registry-server" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.798563 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.803224 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.803524 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sf5hl" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="registry-server" containerID="cri-o://369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326" gracePeriod=30 Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.805205 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bh2kb"] Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.977578 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.977717 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85d55\" (UniqueName: \"kubernetes.io/projected/efe769b7-56d1-4866-bde7-9fc17c66f8cf-kube-api-access-85d55\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:30 crc kubenswrapper[5018]: I1014 06:53:30.977747 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.079295 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.079567 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85d55\" (UniqueName: \"kubernetes.io/projected/efe769b7-56d1-4866-bde7-9fc17c66f8cf-kube-api-access-85d55\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.079585 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.082800 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.086835 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/efe769b7-56d1-4866-bde7-9fc17c66f8cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.095553 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85d55\" (UniqueName: \"kubernetes.io/projected/efe769b7-56d1-4866-bde7-9fc17c66f8cf-kube-api-access-85d55\") pod \"marketplace-operator-79b997595-bh2kb\" (UID: \"efe769b7-56d1-4866-bde7-9fc17c66f8cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.143020 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.159139 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.229408 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.234313 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.244847 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.266344 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.281111 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content\") pod \"f5ace222-4f8c-4058-8497-cb8f12b491fb\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.281187 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities\") pod \"f5ace222-4f8c-4058-8497-cb8f12b491fb\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.281226 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs8rr\" (UniqueName: \"kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr\") pod \"f5ace222-4f8c-4058-8497-cb8f12b491fb\" (UID: \"f5ace222-4f8c-4058-8497-cb8f12b491fb\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.285527 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr" (OuterVolumeSpecName: "kube-api-access-vs8rr") pod "f5ace222-4f8c-4058-8497-cb8f12b491fb" (UID: "f5ace222-4f8c-4058-8497-cb8f12b491fb"). InnerVolumeSpecName "kube-api-access-vs8rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.286370 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities" (OuterVolumeSpecName: "utilities") pod "f5ace222-4f8c-4058-8497-cb8f12b491fb" (UID: "f5ace222-4f8c-4058-8497-cb8f12b491fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.348507 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5ace222-4f8c-4058-8497-cb8f12b491fb" (UID: "f5ace222-4f8c-4058-8497-cb8f12b491fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382288 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities\") pod \"9294a087-ccc1-4955-b730-fd5f3669c190\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382478 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content\") pod \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382503 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities\") pod \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382546 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content\") pod \"9294a087-ccc1-4955-b730-fd5f3669c190\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382583 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca\") pod \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382604 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics\") pod \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382637 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content\") pod \"a3df74a7-f67e-424a-8ac6-69b46c465537\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382680 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmgxp\" (UniqueName: \"kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp\") pod \"a3df74a7-f67e-424a-8ac6-69b46c465537\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382761 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpwfp\" (UniqueName: \"kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp\") pod \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\" (UID: \"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382795 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities\") pod \"a3df74a7-f67e-424a-8ac6-69b46c465537\" (UID: \"a3df74a7-f67e-424a-8ac6-69b46c465537\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382819 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbmvl\" (UniqueName: \"kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl\") pod \"9294a087-ccc1-4955-b730-fd5f3669c190\" (UID: \"9294a087-ccc1-4955-b730-fd5f3669c190\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.382856 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf8bc\" (UniqueName: \"kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc\") pod \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\" (UID: \"b03f9bd8-9597-4358-aae3-4d3cceaf3a47\") " Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383004 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities" (OuterVolumeSpecName: "utilities") pod "9294a087-ccc1-4955-b730-fd5f3669c190" (UID: "9294a087-ccc1-4955-b730-fd5f3669c190"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383119 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383135 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ace222-4f8c-4058-8497-cb8f12b491fb-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383147 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs8rr\" (UniqueName: \"kubernetes.io/projected/f5ace222-4f8c-4058-8497-cb8f12b491fb-kube-api-access-vs8rr\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383160 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383215 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b03f9bd8-9597-4358-aae3-4d3cceaf3a47" (UID: "b03f9bd8-9597-4358-aae3-4d3cceaf3a47"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.383242 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities" (OuterVolumeSpecName: "utilities") pod "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" (UID: "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.384186 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities" (OuterVolumeSpecName: "utilities") pod "a3df74a7-f67e-424a-8ac6-69b46c465537" (UID: "a3df74a7-f67e-424a-8ac6-69b46c465537"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.385665 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp" (OuterVolumeSpecName: "kube-api-access-mmgxp") pod "a3df74a7-f67e-424a-8ac6-69b46c465537" (UID: "a3df74a7-f67e-424a-8ac6-69b46c465537"). InnerVolumeSpecName "kube-api-access-mmgxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.386801 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp" (OuterVolumeSpecName: "kube-api-access-cpwfp") pod "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" (UID: "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3"). InnerVolumeSpecName "kube-api-access-cpwfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.387512 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl" (OuterVolumeSpecName: "kube-api-access-zbmvl") pod "9294a087-ccc1-4955-b730-fd5f3669c190" (UID: "9294a087-ccc1-4955-b730-fd5f3669c190"). InnerVolumeSpecName "kube-api-access-zbmvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.388067 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.388228 5018 generic.go:334] "Generic (PLEG): container finished" podID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerID="8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199" exitCode=0 Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.388322 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" event={"ID":"b03f9bd8-9597-4358-aae3-4d3cceaf3a47","Type":"ContainerDied","Data":"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.388475 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mmmgm" event={"ID":"b03f9bd8-9597-4358-aae3-4d3cceaf3a47","Type":"ContainerDied","Data":"fd4fe0158ddeec38f260290b3845cc5576711b6b4178c3784effd7ccc9c8958c"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.388496 5018 scope.go:117] "RemoveContainer" containerID="8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.389453 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b03f9bd8-9597-4358-aae3-4d3cceaf3a47" (UID: "b03f9bd8-9597-4358-aae3-4d3cceaf3a47"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.389519 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc" (OuterVolumeSpecName: "kube-api-access-zf8bc") pod "b03f9bd8-9597-4358-aae3-4d3cceaf3a47" (UID: "b03f9bd8-9597-4358-aae3-4d3cceaf3a47"). InnerVolumeSpecName "kube-api-access-zf8bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.392100 5018 generic.go:334] "Generic (PLEG): container finished" podID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerID="503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf" exitCode=0 Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.392186 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerDied","Data":"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.392331 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmnmj" event={"ID":"74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3","Type":"ContainerDied","Data":"606cb3b384febea3e0f5562cb8eb42f31dbb923c38d6a4ed6ee66981e89f7fa6"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.392440 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmnmj" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.397134 5018 generic.go:334] "Generic (PLEG): container finished" podID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerID="92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68" exitCode=0 Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.397388 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerDied","Data":"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.397409 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9lcsh" event={"ID":"f5ace222-4f8c-4058-8497-cb8f12b491fb","Type":"ContainerDied","Data":"997a71f2facbe6177ea4554c491ffaf1caa6c5d15db2c9cd591d0465b337feb5"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.397467 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9lcsh" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.402818 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" (UID: "74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.405589 5018 generic.go:334] "Generic (PLEG): container finished" podID="9294a087-ccc1-4955-b730-fd5f3669c190" containerID="369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326" exitCode=0 Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.405669 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerDied","Data":"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.405697 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sf5hl" event={"ID":"9294a087-ccc1-4955-b730-fd5f3669c190","Type":"ContainerDied","Data":"79bb01d2ebd71db655a1647a06c16e0b5e35d0fd8a28bb7769d9f6b6720839be"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.405771 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sf5hl" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.411297 5018 generic.go:334] "Generic (PLEG): container finished" podID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerID="27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4" exitCode=0 Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.411331 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerDied","Data":"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.411357 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9d9j" event={"ID":"a3df74a7-f67e-424a-8ac6-69b46c465537","Type":"ContainerDied","Data":"a9f4488145961e600a13a637906827d6c0da1a0fc293223c5f57d8d2a208cc40"} Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.411378 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9d9j" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.421931 5018 scope.go:117] "RemoveContainer" containerID="8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.422195 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199\": container with ID starting with 8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199 not found: ID does not exist" containerID="8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.422222 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199"} err="failed to get container status \"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199\": rpc error: code = NotFound desc = could not find container \"8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199\": container with ID starting with 8b9961736a5bd6333c6ce6e434910e9da78544e2b6da1d041d5b24f2076dd199 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.422240 5018 scope.go:117] "RemoveContainer" containerID="503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.443683 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.443750 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9lcsh"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.443767 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3df74a7-f67e-424a-8ac6-69b46c465537" (UID: "a3df74a7-f67e-424a-8ac6-69b46c465537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.445440 5018 scope.go:117] "RemoveContainer" containerID="69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.458120 5018 scope.go:117] "RemoveContainer" containerID="d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.470802 5018 scope.go:117] "RemoveContainer" containerID="503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.471221 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf\": container with ID starting with 503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf not found: ID does not exist" containerID="503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471254 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf"} err="failed to get container status \"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf\": rpc error: code = NotFound desc = could not find container \"503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf\": container with ID starting with 503f8f714151f21b43550651f304c788eff1f58dc1cb9f1ab30fe75a616464cf not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471279 5018 scope.go:117] "RemoveContainer" containerID="69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.471510 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b\": container with ID starting with 69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b not found: ID does not exist" containerID="69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471534 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b"} err="failed to get container status \"69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b\": rpc error: code = NotFound desc = could not find container \"69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b\": container with ID starting with 69a33221fcbc27dfe319b2d92ec382669cb0382ca8fa23e89f112d4eab26763b not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471551 5018 scope.go:117] "RemoveContainer" containerID="d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.471784 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc\": container with ID starting with d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc not found: ID does not exist" containerID="d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471811 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc"} err="failed to get container status \"d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc\": rpc error: code = NotFound desc = could not find container \"d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc\": container with ID starting with d7187bc69e21ca4062702a072e6414a764dcc3c7fdbf4eadd559b952cd92e4bc not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.471829 5018 scope.go:117] "RemoveContainer" containerID="92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483828 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483862 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483877 5018 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483888 5018 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483900 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483910 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmgxp\" (UniqueName: \"kubernetes.io/projected/a3df74a7-f67e-424a-8ac6-69b46c465537-kube-api-access-mmgxp\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483922 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpwfp\" (UniqueName: \"kubernetes.io/projected/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3-kube-api-access-cpwfp\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483932 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3df74a7-f67e-424a-8ac6-69b46c465537-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483942 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbmvl\" (UniqueName: \"kubernetes.io/projected/9294a087-ccc1-4955-b730-fd5f3669c190-kube-api-access-zbmvl\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.483952 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf8bc\" (UniqueName: \"kubernetes.io/projected/b03f9bd8-9597-4358-aae3-4d3cceaf3a47-kube-api-access-zf8bc\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.484068 5018 scope.go:117] "RemoveContainer" containerID="bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.500742 5018 scope.go:117] "RemoveContainer" containerID="4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.501246 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9294a087-ccc1-4955-b730-fd5f3669c190" (UID: "9294a087-ccc1-4955-b730-fd5f3669c190"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.513341 5018 scope.go:117] "RemoveContainer" containerID="92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.513691 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68\": container with ID starting with 92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68 not found: ID does not exist" containerID="92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.513716 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68"} err="failed to get container status \"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68\": rpc error: code = NotFound desc = could not find container \"92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68\": container with ID starting with 92f0e547e62c547f7a2fe5fa4f34784d7cc46f74bf014d2419844ce38a8cfe68 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.513734 5018 scope.go:117] "RemoveContainer" containerID="bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.513999 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241\": container with ID starting with bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241 not found: ID does not exist" containerID="bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.514020 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241"} err="failed to get container status \"bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241\": rpc error: code = NotFound desc = could not find container \"bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241\": container with ID starting with bd2eb051fe775c9c8cfbc9c8d42d32dc9251eb190bd5b6b67f2770d780081241 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.514033 5018 scope.go:117] "RemoveContainer" containerID="4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.514254 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6\": container with ID starting with 4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6 not found: ID does not exist" containerID="4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.514274 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6"} err="failed to get container status \"4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6\": rpc error: code = NotFound desc = could not find container \"4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6\": container with ID starting with 4c0050c2f755b6d84a1e6c8e5620acd858b4deeaa74f18661260952f2c390bc6 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.514284 5018 scope.go:117] "RemoveContainer" containerID="369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.531477 5018 scope.go:117] "RemoveContainer" containerID="ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.552332 5018 scope.go:117] "RemoveContainer" containerID="0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.565173 5018 scope.go:117] "RemoveContainer" containerID="369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.565654 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326\": container with ID starting with 369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326 not found: ID does not exist" containerID="369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.565775 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326"} err="failed to get container status \"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326\": rpc error: code = NotFound desc = could not find container \"369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326\": container with ID starting with 369767cba9a129cfd5c0d5f3b0d1802a2b2d469607575d489015295c7782f326 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.565811 5018 scope.go:117] "RemoveContainer" containerID="ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.566252 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31\": container with ID starting with ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31 not found: ID does not exist" containerID="ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.566344 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31"} err="failed to get container status \"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31\": rpc error: code = NotFound desc = could not find container \"ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31\": container with ID starting with ec6bf2bec5eeb9eb866e6e5684b304751621786e44e40261318cd75882b91e31 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.566412 5018 scope.go:117] "RemoveContainer" containerID="0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.566993 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663\": container with ID starting with 0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663 not found: ID does not exist" containerID="0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.567019 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663"} err="failed to get container status \"0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663\": rpc error: code = NotFound desc = could not find container \"0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663\": container with ID starting with 0c7ea7e6417fad719faf67ce190944e48b25f3f671c7e3db1092903f3d219663 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.567038 5018 scope.go:117] "RemoveContainer" containerID="27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.583892 5018 scope.go:117] "RemoveContainer" containerID="bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.584806 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9294a087-ccc1-4955-b730-fd5f3669c190-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.607687 5018 scope.go:117] "RemoveContainer" containerID="8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.612889 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bh2kb"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.625296 5018 scope.go:117] "RemoveContainer" containerID="27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.625690 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4\": container with ID starting with 27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4 not found: ID does not exist" containerID="27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.625734 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4"} err="failed to get container status \"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4\": rpc error: code = NotFound desc = could not find container \"27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4\": container with ID starting with 27b0ba174a7ec341f4ebb2f1a230dd187105d318b8e1c94d03fb9fd0113407a4 not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.625763 5018 scope.go:117] "RemoveContainer" containerID="bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.626082 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c\": container with ID starting with bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c not found: ID does not exist" containerID="bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.626121 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c"} err="failed to get container status \"bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c\": rpc error: code = NotFound desc = could not find container \"bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c\": container with ID starting with bd1841c2a6c1d279a43daf845446f466e8b9906ba8fcb61633ac649e823c621c not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.626150 5018 scope.go:117] "RemoveContainer" containerID="8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe" Oct 14 06:53:31 crc kubenswrapper[5018]: E1014 06:53:31.626410 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe\": container with ID starting with 8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe not found: ID does not exist" containerID="8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.626430 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe"} err="failed to get container status \"8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe\": rpc error: code = NotFound desc = could not find container \"8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe\": container with ID starting with 8875d05a03da69d38f16833f5a8ecc2a0ff637990034439a85b595737dec3bfe not found: ID does not exist" Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.722940 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.734776 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mmmgm"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.738500 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.739527 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sf5hl"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.744404 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.757657 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmnmj"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.765064 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:53:31 crc kubenswrapper[5018]: I1014 06:53:31.769070 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g9d9j"] Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.118872 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.418895 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" event={"ID":"efe769b7-56d1-4866-bde7-9fc17c66f8cf","Type":"ContainerStarted","Data":"945004c945ff248ce17c0d248a622a97bbc4adccb5cd1e3d2f0105d15a6018e9"} Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.419327 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.419344 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" event={"ID":"efe769b7-56d1-4866-bde7-9fc17c66f8cf","Type":"ContainerStarted","Data":"385a71afb352eea0c3b8b784008e440e85011906821f1885c10fc524bfff1a95"} Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.424069 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.459187 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bh2kb" podStartSLOduration=2.4591695700000002 podStartE2EDuration="2.45916957s" podCreationTimestamp="2025-10-14 06:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:53:32.441310293 +0000 UTC m=+229.025356920" watchObservedRunningTime="2025-10-14 06:53:32.45916957 +0000 UTC m=+229.043216197" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.463371 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.463410 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.463442 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.463890 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.463945 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7" gracePeriod=600 Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.618171 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" path="/var/lib/kubelet/pods/74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3/volumes" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.618747 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" path="/var/lib/kubelet/pods/9294a087-ccc1-4955-b730-fd5f3669c190/volumes" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.619270 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" path="/var/lib/kubelet/pods/a3df74a7-f67e-424a-8ac6-69b46c465537/volumes" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.620254 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" path="/var/lib/kubelet/pods/b03f9bd8-9597-4358-aae3-4d3cceaf3a47/volumes" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.620674 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" path="/var/lib/kubelet/pods/f5ace222-4f8c-4058-8497-cb8f12b491fb/volumes" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.970920 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5g2cw"] Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971454 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971469 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971481 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971490 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971503 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971511 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971523 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971531 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971542 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971550 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971560 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971568 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971582 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971590 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971603 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971612 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971627 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971650 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="extract-utilities" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971662 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971670 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="extract-content" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971678 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971685 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971696 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971703 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: E1014 06:53:32.971716 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971724 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971857 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a7c05d-8be0-4bc3-8f60-06d8abdfa9a3" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971872 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ace222-4f8c-4058-8497-cb8f12b491fb" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971883 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9294a087-ccc1-4955-b730-fd5f3669c190" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971891 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03f9bd8-9597-4358-aae3-4d3cceaf3a47" containerName="marketplace-operator" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.971904 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3df74a7-f67e-424a-8ac6-69b46c465537" containerName="registry-server" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.972739 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.975233 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 06:53:32 crc kubenswrapper[5018]: I1014 06:53:32.984324 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5g2cw"] Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.104331 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4szc\" (UniqueName: \"kubernetes.io/projected/77fd3585-d309-4ec2-892a-c255f69032b3-kube-api-access-x4szc\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.104492 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-utilities\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.104583 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-catalog-content\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.171683 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-csprd"] Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.172637 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.174703 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.185676 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csprd"] Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.205467 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-utilities\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.205526 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-catalog-content\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.205598 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4szc\" (UniqueName: \"kubernetes.io/projected/77fd3585-d309-4ec2-892a-c255f69032b3-kube-api-access-x4szc\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.206858 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-utilities\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.206998 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77fd3585-d309-4ec2-892a-c255f69032b3-catalog-content\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.231109 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4szc\" (UniqueName: \"kubernetes.io/projected/77fd3585-d309-4ec2-892a-c255f69032b3-kube-api-access-x4szc\") pod \"certified-operators-5g2cw\" (UID: \"77fd3585-d309-4ec2-892a-c255f69032b3\") " pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.286529 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.307178 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-utilities\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.307262 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrbfn\" (UniqueName: \"kubernetes.io/projected/171f6e90-d807-4e7d-8e5e-20951b92fb7a-kube-api-access-nrbfn\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.307290 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-catalog-content\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.408611 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-utilities\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.408691 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrbfn\" (UniqueName: \"kubernetes.io/projected/171f6e90-d807-4e7d-8e5e-20951b92fb7a-kube-api-access-nrbfn\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.408730 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-catalog-content\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.409187 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-catalog-content\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.409763 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171f6e90-d807-4e7d-8e5e-20951b92fb7a-utilities\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.430036 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7" exitCode=0 Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.430118 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7"} Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.430156 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8"} Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.442841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrbfn\" (UniqueName: \"kubernetes.io/projected/171f6e90-d807-4e7d-8e5e-20951b92fb7a-kube-api-access-nrbfn\") pod \"redhat-marketplace-csprd\" (UID: \"171f6e90-d807-4e7d-8e5e-20951b92fb7a\") " pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.496907 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.735929 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5g2cw"] Oct 14 06:53:33 crc kubenswrapper[5018]: W1014 06:53:33.741140 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77fd3585_d309_4ec2_892a_c255f69032b3.slice/crio-5bf34e596b0137bae83d6ff9a98205508d5519719e04fd1422aa12844f10f99f WatchSource:0}: Error finding container 5bf34e596b0137bae83d6ff9a98205508d5519719e04fd1422aa12844f10f99f: Status 404 returned error can't find the container with id 5bf34e596b0137bae83d6ff9a98205508d5519719e04fd1422aa12844f10f99f Oct 14 06:53:33 crc kubenswrapper[5018]: I1014 06:53:33.876123 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-csprd"] Oct 14 06:53:33 crc kubenswrapper[5018]: W1014 06:53:33.882639 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171f6e90_d807_4e7d_8e5e_20951b92fb7a.slice/crio-df185a6a1316a7bd3373453787bcccdb32212e8d144ecf9b17a62119e1084b15 WatchSource:0}: Error finding container df185a6a1316a7bd3373453787bcccdb32212e8d144ecf9b17a62119e1084b15: Status 404 returned error can't find the container with id df185a6a1316a7bd3373453787bcccdb32212e8d144ecf9b17a62119e1084b15 Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.438385 5018 generic.go:334] "Generic (PLEG): container finished" podID="171f6e90-d807-4e7d-8e5e-20951b92fb7a" containerID="6a9eee564bdd8075e586535f8811ad0682a2c910ef69c26ccf9bfab52d7d806a" exitCode=0 Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.438458 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csprd" event={"ID":"171f6e90-d807-4e7d-8e5e-20951b92fb7a","Type":"ContainerDied","Data":"6a9eee564bdd8075e586535f8811ad0682a2c910ef69c26ccf9bfab52d7d806a"} Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.438831 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csprd" event={"ID":"171f6e90-d807-4e7d-8e5e-20951b92fb7a","Type":"ContainerStarted","Data":"df185a6a1316a7bd3373453787bcccdb32212e8d144ecf9b17a62119e1084b15"} Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.440871 5018 generic.go:334] "Generic (PLEG): container finished" podID="77fd3585-d309-4ec2-892a-c255f69032b3" containerID="cbfc3699498e691b1eb4cda37f3e8f539de637790b48b3858066dcb06a202483" exitCode=0 Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.440949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5g2cw" event={"ID":"77fd3585-d309-4ec2-892a-c255f69032b3","Type":"ContainerDied","Data":"cbfc3699498e691b1eb4cda37f3e8f539de637790b48b3858066dcb06a202483"} Oct 14 06:53:34 crc kubenswrapper[5018]: I1014 06:53:34.440995 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5g2cw" event={"ID":"77fd3585-d309-4ec2-892a-c255f69032b3","Type":"ContainerStarted","Data":"5bf34e596b0137bae83d6ff9a98205508d5519719e04fd1422aa12844f10f99f"} Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.376083 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.378671 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.380631 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.387915 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.446664 5018 generic.go:334] "Generic (PLEG): container finished" podID="171f6e90-d807-4e7d-8e5e-20951b92fb7a" containerID="45634f05515702534e65e1a57c80899dfc351359252fbf38bd26b227befdda12" exitCode=0 Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.446709 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csprd" event={"ID":"171f6e90-d807-4e7d-8e5e-20951b92fb7a","Type":"ContainerDied","Data":"45634f05515702534e65e1a57c80899dfc351359252fbf38bd26b227befdda12"} Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.532660 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ckcr\" (UniqueName: \"kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.532721 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.532750 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.566421 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.567317 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.569563 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.581314 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.634275 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.634314 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635108 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635518 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635819 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635878 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635900 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b965f\" (UniqueName: \"kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.635935 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ckcr\" (UniqueName: \"kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.653435 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ckcr\" (UniqueName: \"kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr\") pod \"community-operators-lvlpg\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.701391 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.737114 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b965f\" (UniqueName: \"kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.737177 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.737201 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.737704 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.737994 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.753397 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b965f\" (UniqueName: \"kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f\") pod \"redhat-operators-db6sr\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:35 crc kubenswrapper[5018]: I1014 06:53:35.878991 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.079288 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 06:53:36 crc kubenswrapper[5018]: W1014 06:53:36.084198 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe2fff80_efbf_4372_a47c_259e5a502e33.slice/crio-26516a7ac17da83a531e73975130dfd12c65a8ee66b7c123de35be8e401b7e30 WatchSource:0}: Error finding container 26516a7ac17da83a531e73975130dfd12c65a8ee66b7c123de35be8e401b7e30: Status 404 returned error can't find the container with id 26516a7ac17da83a531e73975130dfd12c65a8ee66b7c123de35be8e401b7e30 Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.084646 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 06:53:36 crc kubenswrapper[5018]: W1014 06:53:36.095080 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod574592ff_a962_4052_a21f_a114aa3869ff.slice/crio-6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341 WatchSource:0}: Error finding container 6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341: Status 404 returned error can't find the container with id 6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341 Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.455129 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-csprd" event={"ID":"171f6e90-d807-4e7d-8e5e-20951b92fb7a","Type":"ContainerStarted","Data":"e21ab18f9af4769ce8201efad97e271e10c56bf483637668ef59add81da3f609"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.456954 5018 generic.go:334] "Generic (PLEG): container finished" podID="77fd3585-d309-4ec2-892a-c255f69032b3" containerID="0a97b0b9df3e9cf5c0a053f76f637f5ee8f529d8d701b6ceb18240c0a06205ba" exitCode=0 Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.457041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5g2cw" event={"ID":"77fd3585-d309-4ec2-892a-c255f69032b3","Type":"ContainerDied","Data":"0a97b0b9df3e9cf5c0a053f76f637f5ee8f529d8d701b6ceb18240c0a06205ba"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.458466 5018 generic.go:334] "Generic (PLEG): container finished" podID="574592ff-a962-4052-a21f-a114aa3869ff" containerID="c253c6fedb9642a2dde1de157bb555c5a0b393195a9aacda178a94c732244b63" exitCode=0 Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.458934 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerDied","Data":"c253c6fedb9642a2dde1de157bb555c5a0b393195a9aacda178a94c732244b63"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.459115 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerStarted","Data":"6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.460326 5018 generic.go:334] "Generic (PLEG): container finished" podID="be2fff80-efbf-4372-a47c-259e5a502e33" containerID="8ac45eb863dcb797a277ad4ce6b7debb115c1b943fc32c4c9451123e5824e7b9" exitCode=0 Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.460354 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerDied","Data":"8ac45eb863dcb797a277ad4ce6b7debb115c1b943fc32c4c9451123e5824e7b9"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.460375 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerStarted","Data":"26516a7ac17da83a531e73975130dfd12c65a8ee66b7c123de35be8e401b7e30"} Oct 14 06:53:36 crc kubenswrapper[5018]: I1014 06:53:36.479620 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-csprd" podStartSLOduration=2.048513589 podStartE2EDuration="3.479599022s" podCreationTimestamp="2025-10-14 06:53:33 +0000 UTC" firstStartedPulling="2025-10-14 06:53:34.446949547 +0000 UTC m=+231.030996164" lastFinishedPulling="2025-10-14 06:53:35.87803497 +0000 UTC m=+232.462081597" observedRunningTime="2025-10-14 06:53:36.476205204 +0000 UTC m=+233.060251831" watchObservedRunningTime="2025-10-14 06:53:36.479599022 +0000 UTC m=+233.063645649" Oct 14 06:53:37 crc kubenswrapper[5018]: I1014 06:53:37.468747 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5g2cw" event={"ID":"77fd3585-d309-4ec2-892a-c255f69032b3","Type":"ContainerStarted","Data":"1397f3c8c275e723a1e962ddade8e678c7f273c4c371c069fb35bba4efec6052"} Oct 14 06:53:37 crc kubenswrapper[5018]: I1014 06:53:37.472009 5018 generic.go:334] "Generic (PLEG): container finished" podID="574592ff-a962-4052-a21f-a114aa3869ff" containerID="22bd5dc428dee0e158c5723a6542f71c4bc9bb13b14d4921138c22adac5c3224" exitCode=0 Oct 14 06:53:37 crc kubenswrapper[5018]: I1014 06:53:37.472105 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerDied","Data":"22bd5dc428dee0e158c5723a6542f71c4bc9bb13b14d4921138c22adac5c3224"} Oct 14 06:53:37 crc kubenswrapper[5018]: I1014 06:53:37.505584 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5g2cw" podStartSLOduration=2.883268776 podStartE2EDuration="5.505565329s" podCreationTimestamp="2025-10-14 06:53:32 +0000 UTC" firstStartedPulling="2025-10-14 06:53:34.44224543 +0000 UTC m=+231.026292057" lastFinishedPulling="2025-10-14 06:53:37.064541983 +0000 UTC m=+233.648588610" observedRunningTime="2025-10-14 06:53:37.488648889 +0000 UTC m=+234.072695566" watchObservedRunningTime="2025-10-14 06:53:37.505565329 +0000 UTC m=+234.089611956" Oct 14 06:53:38 crc kubenswrapper[5018]: I1014 06:53:38.477391 5018 generic.go:334] "Generic (PLEG): container finished" podID="be2fff80-efbf-4372-a47c-259e5a502e33" containerID="19af5e219d45cf00b4a302a66765b247cd210f21f36e2a2cef71c054cbfb9252" exitCode=0 Oct 14 06:53:38 crc kubenswrapper[5018]: I1014 06:53:38.477817 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerDied","Data":"19af5e219d45cf00b4a302a66765b247cd210f21f36e2a2cef71c054cbfb9252"} Oct 14 06:53:38 crc kubenswrapper[5018]: I1014 06:53:38.481313 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerStarted","Data":"6f7e5b36772fb725cf1c59b62ec20d30f1e1bd0e3b55ca2af90ddaabd078aa8d"} Oct 14 06:53:38 crc kubenswrapper[5018]: I1014 06:53:38.512075 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lvlpg" podStartSLOduration=2.065982635 podStartE2EDuration="3.512058403s" podCreationTimestamp="2025-10-14 06:53:35 +0000 UTC" firstStartedPulling="2025-10-14 06:53:36.459572882 +0000 UTC m=+233.043619529" lastFinishedPulling="2025-10-14 06:53:37.90564867 +0000 UTC m=+234.489695297" observedRunningTime="2025-10-14 06:53:38.509041955 +0000 UTC m=+235.093088582" watchObservedRunningTime="2025-10-14 06:53:38.512058403 +0000 UTC m=+235.096105030" Oct 14 06:53:39 crc kubenswrapper[5018]: I1014 06:53:39.501709 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerStarted","Data":"d988f06ec7750d8716be5da2d3fd09f55f000182661cb82375fd7e5084a03133"} Oct 14 06:53:39 crc kubenswrapper[5018]: I1014 06:53:39.523211 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-db6sr" podStartSLOduration=2.002453435 podStartE2EDuration="4.52319245s" podCreationTimestamp="2025-10-14 06:53:35 +0000 UTC" firstStartedPulling="2025-10-14 06:53:36.462343932 +0000 UTC m=+233.046390559" lastFinishedPulling="2025-10-14 06:53:38.983082947 +0000 UTC m=+235.567129574" observedRunningTime="2025-10-14 06:53:39.520328978 +0000 UTC m=+236.104375605" watchObservedRunningTime="2025-10-14 06:53:39.52319245 +0000 UTC m=+236.107239077" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.286845 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.288126 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.341975 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.497670 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.497726 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.537613 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.561789 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5g2cw" Oct 14 06:53:43 crc kubenswrapper[5018]: I1014 06:53:43.575805 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-csprd" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.701905 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.702415 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.755600 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.879683 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.879788 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:45 crc kubenswrapper[5018]: I1014 06:53:45.923909 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:46 crc kubenswrapper[5018]: I1014 06:53:46.618453 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 06:53:46 crc kubenswrapper[5018]: I1014 06:53:46.627316 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.147502 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" podUID="aaa522ed-38fb-40f0-9015-b97956137bc8" containerName="oauth-openshift" containerID="cri-o://392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9" gracePeriod=15 Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.597225 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.635053 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-9nd57"] Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.638411 5018 generic.go:334] "Generic (PLEG): container finished" podID="aaa522ed-38fb-40f0-9015-b97956137bc8" containerID="392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9" exitCode=0 Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.638496 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" Oct 14 06:53:57 crc kubenswrapper[5018]: E1014 06:53:57.647277 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa522ed-38fb-40f0-9015-b97956137bc8" containerName="oauth-openshift" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647305 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa522ed-38fb-40f0-9015-b97956137bc8" containerName="oauth-openshift" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647448 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa522ed-38fb-40f0-9015-b97956137bc8" containerName="oauth-openshift" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" event={"ID":"aaa522ed-38fb-40f0-9015-b97956137bc8","Type":"ContainerDied","Data":"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9"} Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647823 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-htrtd" event={"ID":"aaa522ed-38fb-40f0-9015-b97956137bc8","Type":"ContainerDied","Data":"0372713277d9b65db71add3414ab56be57f6e7f86199591b8c0dfe6862dba6c5"} Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647840 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-9nd57"] Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.647860 5018 scope.go:117] "RemoveContainer" containerID="392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.648020 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.675204 5018 scope.go:117] "RemoveContainer" containerID="392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9" Oct 14 06:53:57 crc kubenswrapper[5018]: E1014 06:53:57.675997 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9\": container with ID starting with 392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9 not found: ID does not exist" containerID="392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.676063 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9"} err="failed to get container status \"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9\": rpc error: code = NotFound desc = could not find container \"392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9\": container with ID starting with 392d6939908cc09f93927094af4f990171566be275b209b8ac9169f637704ec9 not found: ID does not exist" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.711783 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.711848 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.711910 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.711946 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.711978 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712013 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712045 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712093 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712126 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712172 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712228 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712258 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712314 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68bcd\" (UniqueName: \"kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712363 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session\") pod \"aaa522ed-38fb-40f0-9015-b97956137bc8\" (UID: \"aaa522ed-38fb-40f0-9015-b97956137bc8\") " Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.712539 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.713049 5018 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.713094 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.713215 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.713238 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.713873 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.733163 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.733183 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd" (OuterVolumeSpecName: "kube-api-access-68bcd") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "kube-api-access-68bcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.733664 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.733789 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.734009 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.734442 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.734566 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.734751 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.735788 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "aaa522ed-38fb-40f0-9015-b97956137bc8" (UID: "aaa522ed-38fb-40f0-9015-b97956137bc8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.814990 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815059 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815115 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815155 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815195 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815234 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px7n8\" (UniqueName: \"kubernetes.io/projected/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-kube-api-access-px7n8\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815268 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-dir\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815298 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815347 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-policies\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815377 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815442 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815476 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815511 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.815896 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816013 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816037 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816057 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68bcd\" (UniqueName: \"kubernetes.io/projected/aaa522ed-38fb-40f0-9015-b97956137bc8-kube-api-access-68bcd\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816077 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816095 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816115 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816134 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816153 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816171 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816190 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816209 5018 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816227 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.816248 5018 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa522ed-38fb-40f0-9015-b97956137bc8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917519 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917586 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917669 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917721 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917764 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917795 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917842 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917884 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.917960 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.918021 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px7n8\" (UniqueName: \"kubernetes.io/projected/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-kube-api-access-px7n8\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.918055 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-dir\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.918089 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.918138 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-policies\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.918169 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.919165 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-dir\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.919177 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.920210 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.920470 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-audit-policies\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.921119 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.925004 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.925341 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.925709 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.926396 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.927822 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.927901 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.928067 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.930360 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.947695 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px7n8\" (UniqueName: \"kubernetes.io/projected/9e7aec5f-1acb-4984-ac93-6bc2304c18b6-kube-api-access-px7n8\") pod \"oauth-openshift-5555647bc4-9nd57\" (UID: \"9e7aec5f-1acb-4984-ac93-6bc2304c18b6\") " pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.966919 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.968343 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:57 crc kubenswrapper[5018]: I1014 06:53:57.979793 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-htrtd"] Oct 14 06:53:58 crc kubenswrapper[5018]: I1014 06:53:58.440225 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-9nd57"] Oct 14 06:53:58 crc kubenswrapper[5018]: I1014 06:53:58.611556 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaa522ed-38fb-40f0-9015-b97956137bc8" path="/var/lib/kubelet/pods/aaa522ed-38fb-40f0-9015-b97956137bc8/volumes" Oct 14 06:53:58 crc kubenswrapper[5018]: I1014 06:53:58.650699 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" event={"ID":"9e7aec5f-1acb-4984-ac93-6bc2304c18b6","Type":"ContainerStarted","Data":"0497e98df75d5396aa66f8660b9d5e7ede5c6c25737343d7f1b0e604e31e5ab8"} Oct 14 06:53:59 crc kubenswrapper[5018]: I1014 06:53:59.661327 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" event={"ID":"9e7aec5f-1acb-4984-ac93-6bc2304c18b6","Type":"ContainerStarted","Data":"6cae036de62abdb4cf93dba97fd52dd87f9ad11a53d5c965d9a5db770816c71c"} Oct 14 06:53:59 crc kubenswrapper[5018]: I1014 06:53:59.661884 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:59 crc kubenswrapper[5018]: I1014 06:53:59.677060 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" Oct 14 06:53:59 crc kubenswrapper[5018]: I1014 06:53:59.707899 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5555647bc4-9nd57" podStartSLOduration=27.707870572 podStartE2EDuration="27.707870572s" podCreationTimestamp="2025-10-14 06:53:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:53:59.699003736 +0000 UTC m=+256.283086414" watchObservedRunningTime="2025-10-14 06:53:59.707870572 +0000 UTC m=+256.291917239" Oct 14 06:55:32 crc kubenswrapper[5018]: I1014 06:55:32.464098 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:55:32 crc kubenswrapper[5018]: I1014 06:55:32.466493 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:56:02 crc kubenswrapper[5018]: I1014 06:56:02.463577 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:56:02 crc kubenswrapper[5018]: I1014 06:56:02.464196 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.463419 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.464098 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.464168 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.465115 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.465292 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8" gracePeriod=600 Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.648102 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8" exitCode=0 Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.648158 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8"} Oct 14 06:56:32 crc kubenswrapper[5018]: I1014 06:56:32.648209 5018 scope.go:117] "RemoveContainer" containerID="376c16304f3090be338357cf04d7e564530e2dafd70877c36c39fdf2d9b67af7" Oct 14 06:56:33 crc kubenswrapper[5018]: I1014 06:56:33.659239 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39"} Oct 14 06:58:05 crc kubenswrapper[5018]: I1014 06:58:05.986537 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hdfsn"] Oct 14 06:58:05 crc kubenswrapper[5018]: I1014 06:58:05.987806 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.007169 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hdfsn"] Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172417 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-certificates\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172707 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172728 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vgrk\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-kube-api-access-4vgrk\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172745 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-trusted-ca\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172772 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172790 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-bound-sa-token\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172810 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.172854 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-tls\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.202877 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274204 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-certificates\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274251 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274277 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vgrk\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-kube-api-access-4vgrk\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274303 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-trusted-ca\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274332 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-bound-sa-token\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274357 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274426 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-tls\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.274917 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.276475 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-certificates\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.276809 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-trusted-ca\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.282425 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-registry-tls\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.284007 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.292481 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vgrk\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-kube-api-access-4vgrk\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.298067 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0d3462d-4d12-4591-b8b4-5d84e7c514b6-bound-sa-token\") pod \"image-registry-66df7c8f76-hdfsn\" (UID: \"d0d3462d-4d12-4591-b8b4-5d84e7c514b6\") " pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.307648 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:06 crc kubenswrapper[5018]: I1014 06:58:06.613663 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hdfsn"] Oct 14 06:58:07 crc kubenswrapper[5018]: I1014 06:58:07.278523 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" event={"ID":"d0d3462d-4d12-4591-b8b4-5d84e7c514b6","Type":"ContainerStarted","Data":"ff8abd9a0830e55826c8368c605ac07509b83787b84ae6aea311a33e53c57d99"} Oct 14 06:58:07 crc kubenswrapper[5018]: I1014 06:58:07.278565 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" event={"ID":"d0d3462d-4d12-4591-b8b4-5d84e7c514b6","Type":"ContainerStarted","Data":"d4e333770f9d37ef9d9adf91a2712885977a893c8b14dd9be50d0a49676ac70c"} Oct 14 06:58:07 crc kubenswrapper[5018]: I1014 06:58:07.278728 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:07 crc kubenswrapper[5018]: I1014 06:58:07.302369 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" podStartSLOduration=2.302344841 podStartE2EDuration="2.302344841s" podCreationTimestamp="2025-10-14 06:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 06:58:07.300497768 +0000 UTC m=+503.884544405" watchObservedRunningTime="2025-10-14 06:58:07.302344841 +0000 UTC m=+503.886391498" Oct 14 06:58:26 crc kubenswrapper[5018]: I1014 06:58:26.315862 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hdfsn" Oct 14 06:58:26 crc kubenswrapper[5018]: I1014 06:58:26.377112 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:58:32 crc kubenswrapper[5018]: I1014 06:58:32.463611 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:58:32 crc kubenswrapper[5018]: I1014 06:58:32.464028 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.440314 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" podUID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" containerName="registry" containerID="cri-o://00a6ac6361dccbf9602f5d7c710ae75d987a35d1a6f3e47982125b0cc1f6305e" gracePeriod=30 Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.628981 5018 generic.go:334] "Generic (PLEG): container finished" podID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" containerID="00a6ac6361dccbf9602f5d7c710ae75d987a35d1a6f3e47982125b0cc1f6305e" exitCode=0 Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.629057 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" event={"ID":"7c6accb3-b364-4b05-97b7-ac8cdcb8db31","Type":"ContainerDied","Data":"00a6ac6361dccbf9602f5d7c710ae75d987a35d1a6f3e47982125b0cc1f6305e"} Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.838938 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937357 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937454 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4dg5\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937528 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937554 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937608 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937806 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937846 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.937875 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets\") pod \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\" (UID: \"7c6accb3-b364-4b05-97b7-ac8cdcb8db31\") " Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.943298 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.949726 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.956696 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.961136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.961144 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.966811 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.966883 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5" (OuterVolumeSpecName: "kube-api-access-q4dg5") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "kube-api-access-q4dg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 06:58:51 crc kubenswrapper[5018]: I1014 06:58:51.988069 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7c6accb3-b364-4b05-97b7-ac8cdcb8db31" (UID: "7c6accb3-b364-4b05-97b7-ac8cdcb8db31"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039296 5018 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039377 5018 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039398 5018 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039418 5018 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039434 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4dg5\" (UniqueName: \"kubernetes.io/projected/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-kube-api-access-q4dg5\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039449 5018 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.039467 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c6accb3-b364-4b05-97b7-ac8cdcb8db31-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.639547 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" event={"ID":"7c6accb3-b364-4b05-97b7-ac8cdcb8db31","Type":"ContainerDied","Data":"7a0dfbfa84b8b327a8c248f2d3a60ae2dc57fc8faa7daaba23ec8f9f315d89f5"} Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.639777 5018 scope.go:117] "RemoveContainer" containerID="00a6ac6361dccbf9602f5d7c710ae75d987a35d1a6f3e47982125b0cc1f6305e" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.639934 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hkvfr" Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.675066 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:58:52 crc kubenswrapper[5018]: I1014 06:58:52.682079 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hkvfr"] Oct 14 06:58:54 crc kubenswrapper[5018]: I1014 06:58:54.619041 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" path="/var/lib/kubelet/pods/7c6accb3-b364-4b05-97b7-ac8cdcb8db31/volumes" Oct 14 06:59:02 crc kubenswrapper[5018]: I1014 06:59:02.463616 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:59:02 crc kubenswrapper[5018]: I1014 06:59:02.463985 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.464169 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.464852 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.464921 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.465739 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.465842 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39" gracePeriod=600 Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.931849 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39" exitCode=0 Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.932601 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39"} Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.932665 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345"} Oct 14 06:59:32 crc kubenswrapper[5018]: I1014 06:59:32.932689 5018 scope.go:117] "RemoveContainer" containerID="c9fa0ba14a85ea038e8ffdf450708a2f4ae4262e7146c3adbef03bf924afd1e8" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.156156 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb"] Oct 14 07:00:00 crc kubenswrapper[5018]: E1014 07:00:00.157336 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" containerName="registry" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.157369 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" containerName="registry" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.157677 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6accb3-b364-4b05-97b7-ac8cdcb8db31" containerName="registry" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.158492 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.164474 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.165952 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.170187 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb"] Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.245035 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.245101 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.245231 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88g52\" (UniqueName: \"kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.346487 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88g52\" (UniqueName: \"kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.346589 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.346708 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.347992 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.356449 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.376694 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88g52\" (UniqueName: \"kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52\") pod \"collect-profiles-29340420-kgksb\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.484509 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:00 crc kubenswrapper[5018]: I1014 07:00:00.716258 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb"] Oct 14 07:00:01 crc kubenswrapper[5018]: I1014 07:00:01.126277 5018 generic.go:334] "Generic (PLEG): container finished" podID="4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" containerID="fc272a2f09dca504610589f45b601b615ee3378941d217d7fe9469fa82e9d158" exitCode=0 Oct 14 07:00:01 crc kubenswrapper[5018]: I1014 07:00:01.126363 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" event={"ID":"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc","Type":"ContainerDied","Data":"fc272a2f09dca504610589f45b601b615ee3378941d217d7fe9469fa82e9d158"} Oct 14 07:00:01 crc kubenswrapper[5018]: I1014 07:00:01.126634 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" event={"ID":"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc","Type":"ContainerStarted","Data":"db994148a076ece07d918bb7623f54a76bdd4012226931eb73e7cac8997a5371"} Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.411450 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.472585 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume\") pod \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.472707 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88g52\" (UniqueName: \"kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52\") pod \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.472768 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume\") pod \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\" (UID: \"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc\") " Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.474146 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" (UID: "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.483855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" (UID: "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.483847 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52" (OuterVolumeSpecName: "kube-api-access-88g52") pod "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" (UID: "4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc"). InnerVolumeSpecName "kube-api-access-88g52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.573776 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.573862 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:00:02 crc kubenswrapper[5018]: I1014 07:00:02.573884 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88g52\" (UniqueName: \"kubernetes.io/projected/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc-kube-api-access-88g52\") on node \"crc\" DevicePath \"\"" Oct 14 07:00:03 crc kubenswrapper[5018]: I1014 07:00:03.141224 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" Oct 14 07:00:03 crc kubenswrapper[5018]: I1014 07:00:03.142707 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb" event={"ID":"4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc","Type":"ContainerDied","Data":"db994148a076ece07d918bb7623f54a76bdd4012226931eb73e7cac8997a5371"} Oct 14 07:00:03 crc kubenswrapper[5018]: I1014 07:00:03.142812 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db994148a076ece07d918bb7623f54a76bdd4012226931eb73e7cac8997a5371" Oct 14 07:01:32 crc kubenswrapper[5018]: I1014 07:01:32.463371 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:01:32 crc kubenswrapper[5018]: I1014 07:01:32.463989 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.042404 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c2zpf"] Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043264 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-controller" containerID="cri-o://c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043397 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="sbdb" containerID="cri-o://c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043627 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="nbdb" containerID="cri-o://ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043679 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="northd" containerID="cri-o://bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043706 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043731 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-node" containerID="cri-o://84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.043759 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-acl-logging" containerID="cri-o://ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.093175 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" containerID="cri-o://a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" gracePeriod=30 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.381620 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/3.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.386443 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovn-acl-logging/0.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.387372 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovn-controller/0.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.388182 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.458573 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ddbv7"] Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.458900 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kubecfg-setup" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.458929 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kubecfg-setup" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.458947 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.458961 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.458975 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.458989 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459003 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-acl-logging" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459017 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-acl-logging" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459036 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459048 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459063 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-node" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459075 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-node" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459095 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="sbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459108 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="sbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459130 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459142 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459155 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459168 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459186 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" containerName="collect-profiles" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459198 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" containerName="collect-profiles" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459212 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="northd" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459224 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="northd" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459240 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="nbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459251 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="nbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459405 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="sbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459425 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="nbdb" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459443 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="northd" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459461 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459477 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-node" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459490 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" containerName="collect-profiles" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459509 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459525 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459542 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459557 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459574 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovn-acl-logging" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459590 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459794 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459809 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.459828 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459839 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.459991 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerName="ovnkube-controller" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.462670 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.526814 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.526886 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.526920 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.526955 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.526997 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527029 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527061 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527092 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527125 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527161 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527190 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527232 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527271 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn5tp\" (UniqueName: \"kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527298 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527330 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527363 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527392 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527422 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527455 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527511 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch\") pod \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\" (UID: \"99edde9a-4116-4bdf-84ff-863e9e6ed55e\") " Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527772 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-ovn\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527814 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-script-lib\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527859 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-var-lib-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527905 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-log-socket\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527961 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.527992 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-env-overrides\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528026 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-kubelet\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528054 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-config\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528087 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-systemd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528123 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-netns\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528161 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/847d2023-88e3-4116-af02-621ce12380a2-ovn-node-metrics-cert\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528155 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528193 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-bin\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528226 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4t8h\" (UniqueName: \"kubernetes.io/projected/847d2023-88e3-4116-af02-621ce12380a2-kube-api-access-p4t8h\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528258 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-systemd-units\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528286 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-slash\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528329 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528362 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-node-log\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528391 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-netd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528422 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-etc-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528465 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528314 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log" (OuterVolumeSpecName: "node-log") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528293 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528367 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528345 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528397 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528367 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528398 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket" (OuterVolumeSpecName: "log-socket") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528416 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528432 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528439 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528452 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash" (OuterVolumeSpecName: "host-slash") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528476 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528703 5018 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.528523 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.529050 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.529199 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.529708 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.533555 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.533817 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp" (OuterVolumeSpecName: "kube-api-access-zn5tp") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "kube-api-access-zn5tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.553116 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "99edde9a-4116-4bdf-84ff-863e9e6ed55e" (UID: "99edde9a-4116-4bdf-84ff-863e9e6ed55e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629561 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4t8h\" (UniqueName: \"kubernetes.io/projected/847d2023-88e3-4116-af02-621ce12380a2-kube-api-access-p4t8h\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629624 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-systemd-units\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629681 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-slash\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629717 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629821 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-slash\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629825 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-systemd-units\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629881 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-node-log\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629913 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-node-log\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629911 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-netd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629944 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630028 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-etc-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.629965 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-etc-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630081 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-netd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630143 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630113 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630236 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-ovn\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630279 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-script-lib\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630355 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-var-lib-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630424 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-log-socket\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630421 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-ovn\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630520 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-log-socket\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630460 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-var-lib-openvswitch\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630547 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630587 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630591 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-env-overrides\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630712 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-config\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-kubelet\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630792 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-systemd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630837 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-netns\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630876 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/847d2023-88e3-4116-af02-621ce12380a2-ovn-node-metrics-cert\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630914 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-bin\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.630990 5018 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-log-socket\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631009 5018 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-node-log\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631028 5018 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631046 5018 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631065 5018 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631081 5018 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-slash\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631100 5018 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631117 5018 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631139 5018 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631158 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn5tp\" (UniqueName: \"kubernetes.io/projected/99edde9a-4116-4bdf-84ff-863e9e6ed55e-kube-api-access-zn5tp\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631176 5018 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631172 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-run-netns\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-cni-bin\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631076 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-host-kubelet\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631192 5018 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631326 5018 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631356 5018 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631383 5018 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631108 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/847d2023-88e3-4116-af02-621ce12380a2-run-systemd\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631411 5018 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631499 5018 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631529 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/99edde9a-4116-4bdf-84ff-863e9e6ed55e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631564 5018 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/99edde9a-4116-4bdf-84ff-863e9e6ed55e-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631413 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-env-overrides\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.631543 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-script-lib\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.632830 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/847d2023-88e3-4116-af02-621ce12380a2-ovnkube-config\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.635978 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/847d2023-88e3-4116-af02-621ce12380a2-ovn-node-metrics-cert\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.660404 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4t8h\" (UniqueName: \"kubernetes.io/projected/847d2023-88e3-4116-af02-621ce12380a2-kube-api-access-p4t8h\") pod \"ovnkube-node-ddbv7\" (UID: \"847d2023-88e3-4116-af02-621ce12380a2\") " pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.776766 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.826281 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/2.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.827568 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/1.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.827689 5018 generic.go:334] "Generic (PLEG): container finished" podID="f1aa622b-96ac-42aa-8bd1-8fa174e65602" containerID="0eb191c53d3c98d15cfe80fae0123a7b7d712dc828106085b6059d32258a3d98" exitCode=2 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.827749 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerDied","Data":"0eb191c53d3c98d15cfe80fae0123a7b7d712dc828106085b6059d32258a3d98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.827822 5018 scope.go:117] "RemoveContainer" containerID="7a8ba2357c0eeead9138b9733cba48746f53aaa847f1cc2a3c548e468804a24a" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.828784 5018 scope.go:117] "RemoveContainer" containerID="0eb191c53d3c98d15cfe80fae0123a7b7d712dc828106085b6059d32258a3d98" Oct 14 07:01:40 crc kubenswrapper[5018]: E1014 07:01:40.829178 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-r7ftx_openshift-multus(f1aa622b-96ac-42aa-8bd1-8fa174e65602)\"" pod="openshift-multus/multus-r7ftx" podUID="f1aa622b-96ac-42aa-8bd1-8fa174e65602" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.839040 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"eb3d16b6483c35a3764eaeb403b1d305cd764dc9473a8cb4d2aa3485a531300f"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.854727 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovnkube-controller/3.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.872718 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovn-acl-logging/0.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.873907 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c2zpf_99edde9a-4116-4bdf-84ff-863e9e6ed55e/ovn-controller/0.log" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875075 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875124 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875140 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875155 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875171 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875188 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" exitCode=0 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875202 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" exitCode=143 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875219 5018 generic.go:334] "Generic (PLEG): container finished" podID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" exitCode=143 Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875257 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875313 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875309 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875452 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875474 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875495 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875516 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875535 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875548 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875560 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875572 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875584 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875596 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875607 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875647 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875659 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875675 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875693 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875705 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875718 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875729 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875741 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875752 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875764 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875775 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875786 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875797 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875815 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875832 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875844 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875857 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875869 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875883 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875894 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875906 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875917 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875929 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875941 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875957 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2zpf" event={"ID":"99edde9a-4116-4bdf-84ff-863e9e6ed55e","Type":"ContainerDied","Data":"f8b8d9748c40989728b53957cb281152cf9134431fc51993c123e5594ee19121"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.875988 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876001 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876013 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876025 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876037 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876051 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876062 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876073 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876084 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.876095 5018 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.885439 5018 scope.go:117] "RemoveContainer" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.906841 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c2zpf"] Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.910817 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c2zpf"] Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.940261 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.966134 5018 scope.go:117] "RemoveContainer" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:40 crc kubenswrapper[5018]: I1014 07:01:40.989269 5018 scope.go:117] "RemoveContainer" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.068239 5018 scope.go:117] "RemoveContainer" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.088869 5018 scope.go:117] "RemoveContainer" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.103283 5018 scope.go:117] "RemoveContainer" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.123367 5018 scope.go:117] "RemoveContainer" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.142760 5018 scope.go:117] "RemoveContainer" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.164568 5018 scope.go:117] "RemoveContainer" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.228284 5018 scope.go:117] "RemoveContainer" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.228831 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": container with ID starting with a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086 not found: ID does not exist" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.228876 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} err="failed to get container status \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": rpc error: code = NotFound desc = could not find container \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": container with ID starting with a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.228904 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.229340 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": container with ID starting with 5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c not found: ID does not exist" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.229400 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} err="failed to get container status \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": rpc error: code = NotFound desc = could not find container \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": container with ID starting with 5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.229438 5018 scope.go:117] "RemoveContainer" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.229871 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": container with ID starting with c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523 not found: ID does not exist" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.229900 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} err="failed to get container status \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": rpc error: code = NotFound desc = could not find container \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": container with ID starting with c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.229916 5018 scope.go:117] "RemoveContainer" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.230197 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": container with ID starting with ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98 not found: ID does not exist" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.230240 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} err="failed to get container status \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": rpc error: code = NotFound desc = could not find container \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": container with ID starting with ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.230267 5018 scope.go:117] "RemoveContainer" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.230766 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": container with ID starting with bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49 not found: ID does not exist" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.230859 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} err="failed to get container status \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": rpc error: code = NotFound desc = could not find container \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": container with ID starting with bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.230914 5018 scope.go:117] "RemoveContainer" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.231319 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": container with ID starting with 13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4 not found: ID does not exist" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231347 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} err="failed to get container status \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": rpc error: code = NotFound desc = could not find container \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": container with ID starting with 13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231362 5018 scope.go:117] "RemoveContainer" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.231618 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": container with ID starting with 84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2 not found: ID does not exist" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231670 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} err="failed to get container status \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": rpc error: code = NotFound desc = could not find container \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": container with ID starting with 84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231690 5018 scope.go:117] "RemoveContainer" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.231942 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": container with ID starting with ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0 not found: ID does not exist" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231963 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} err="failed to get container status \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": rpc error: code = NotFound desc = could not find container \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": container with ID starting with ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.231975 5018 scope.go:117] "RemoveContainer" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.232270 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": container with ID starting with c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013 not found: ID does not exist" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.232291 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} err="failed to get container status \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": rpc error: code = NotFound desc = could not find container \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": container with ID starting with c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.232303 5018 scope.go:117] "RemoveContainer" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: E1014 07:01:41.232585 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": container with ID starting with 8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f not found: ID does not exist" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.232612 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} err="failed to get container status \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": rpc error: code = NotFound desc = could not find container \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": container with ID starting with 8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.232645 5018 scope.go:117] "RemoveContainer" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.232969 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} err="failed to get container status \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": rpc error: code = NotFound desc = could not find container \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": container with ID starting with a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.233019 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.233322 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} err="failed to get container status \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": rpc error: code = NotFound desc = could not find container \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": container with ID starting with 5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.233364 5018 scope.go:117] "RemoveContainer" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.233618 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} err="failed to get container status \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": rpc error: code = NotFound desc = could not find container \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": container with ID starting with c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.233661 5018 scope.go:117] "RemoveContainer" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234047 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} err="failed to get container status \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": rpc error: code = NotFound desc = could not find container \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": container with ID starting with ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234070 5018 scope.go:117] "RemoveContainer" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234313 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} err="failed to get container status \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": rpc error: code = NotFound desc = could not find container \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": container with ID starting with bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234352 5018 scope.go:117] "RemoveContainer" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234791 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} err="failed to get container status \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": rpc error: code = NotFound desc = could not find container \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": container with ID starting with 13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.234812 5018 scope.go:117] "RemoveContainer" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235086 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} err="failed to get container status \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": rpc error: code = NotFound desc = could not find container \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": container with ID starting with 84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235137 5018 scope.go:117] "RemoveContainer" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235420 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} err="failed to get container status \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": rpc error: code = NotFound desc = could not find container \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": container with ID starting with ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235442 5018 scope.go:117] "RemoveContainer" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235727 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} err="failed to get container status \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": rpc error: code = NotFound desc = could not find container \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": container with ID starting with c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.235767 5018 scope.go:117] "RemoveContainer" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236003 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} err="failed to get container status \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": rpc error: code = NotFound desc = could not find container \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": container with ID starting with 8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236022 5018 scope.go:117] "RemoveContainer" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236249 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} err="failed to get container status \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": rpc error: code = NotFound desc = could not find container \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": container with ID starting with a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236278 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236497 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} err="failed to get container status \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": rpc error: code = NotFound desc = could not find container \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": container with ID starting with 5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236518 5018 scope.go:117] "RemoveContainer" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236848 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} err="failed to get container status \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": rpc error: code = NotFound desc = could not find container \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": container with ID starting with c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.236899 5018 scope.go:117] "RemoveContainer" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237171 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} err="failed to get container status \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": rpc error: code = NotFound desc = could not find container \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": container with ID starting with ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237208 5018 scope.go:117] "RemoveContainer" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237440 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} err="failed to get container status \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": rpc error: code = NotFound desc = could not find container \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": container with ID starting with bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237460 5018 scope.go:117] "RemoveContainer" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237692 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} err="failed to get container status \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": rpc error: code = NotFound desc = could not find container \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": container with ID starting with 13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237721 5018 scope.go:117] "RemoveContainer" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237943 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} err="failed to get container status \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": rpc error: code = NotFound desc = could not find container \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": container with ID starting with 84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.237963 5018 scope.go:117] "RemoveContainer" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238206 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} err="failed to get container status \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": rpc error: code = NotFound desc = could not find container \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": container with ID starting with ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238248 5018 scope.go:117] "RemoveContainer" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238532 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} err="failed to get container status \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": rpc error: code = NotFound desc = could not find container \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": container with ID starting with c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238552 5018 scope.go:117] "RemoveContainer" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238803 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} err="failed to get container status \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": rpc error: code = NotFound desc = could not find container \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": container with ID starting with 8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.238840 5018 scope.go:117] "RemoveContainer" containerID="a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239055 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086"} err="failed to get container status \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": rpc error: code = NotFound desc = could not find container \"a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086\": container with ID starting with a97e9dd2b3b964789f2e1ad96cb2e75cb626cf4424db31a397cc247a644be086 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239073 5018 scope.go:117] "RemoveContainer" containerID="5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239344 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c"} err="failed to get container status \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": rpc error: code = NotFound desc = could not find container \"5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c\": container with ID starting with 5ccae68db460c41aa70bbcd6f3dbf49c5a4b7e2ddcac36c4c078f57dab92fb1c not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239379 5018 scope.go:117] "RemoveContainer" containerID="c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239667 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523"} err="failed to get container status \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": rpc error: code = NotFound desc = could not find container \"c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523\": container with ID starting with c1146f877346714eec2d789ba5b84bc62a2c7aafe576794e3d7cf74eac50c523 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239710 5018 scope.go:117] "RemoveContainer" containerID="ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239957 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98"} err="failed to get container status \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": rpc error: code = NotFound desc = could not find container \"ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98\": container with ID starting with ad6a48d59984f8c0747f37802e462056b833eac258c7f3de4f0ba7686011ea98 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.239982 5018 scope.go:117] "RemoveContainer" containerID="bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240220 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49"} err="failed to get container status \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": rpc error: code = NotFound desc = could not find container \"bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49\": container with ID starting with bf3c71d75f6fdf1702f39ebb131f9c19a1d1674f528a740d95e505a706fbda49 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240254 5018 scope.go:117] "RemoveContainer" containerID="13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240518 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4"} err="failed to get container status \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": rpc error: code = NotFound desc = could not find container \"13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4\": container with ID starting with 13b0b11f220d66136dceb33118d84001644b57c7a4564a4f2d671fb882f04ee4 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240558 5018 scope.go:117] "RemoveContainer" containerID="84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240832 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2"} err="failed to get container status \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": rpc error: code = NotFound desc = could not find container \"84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2\": container with ID starting with 84407a167ba63dc0d20165df979436db7021e9fe51f4c0daed3fe49cfd9141e2 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.240852 5018 scope.go:117] "RemoveContainer" containerID="ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.241077 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0"} err="failed to get container status \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": rpc error: code = NotFound desc = could not find container \"ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0\": container with ID starting with ac39580eb69e174428f1741b31aa4f7dc26790ea628d809c6a06b569a9d46dd0 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.241112 5018 scope.go:117] "RemoveContainer" containerID="c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.241327 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013"} err="failed to get container status \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": rpc error: code = NotFound desc = could not find container \"c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013\": container with ID starting with c5db57c1bfff173387b165fb7fb4841933b570dc0aa2ee89703d728286387013 not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.241347 5018 scope.go:117] "RemoveContainer" containerID="8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.241523 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f"} err="failed to get container status \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": rpc error: code = NotFound desc = could not find container \"8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f\": container with ID starting with 8a7075b3c29e7be0260521b0eabb7da5232fc8872357ed21042fde8ee2f3b53f not found: ID does not exist" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.884509 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/2.log" Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.887699 5018 generic.go:334] "Generic (PLEG): container finished" podID="847d2023-88e3-4116-af02-621ce12380a2" containerID="5799aa1d68f43f38935f4ea8e2f0276c5b71e0a9fe4eae041df29ae95f675905" exitCode=0 Oct 14 07:01:41 crc kubenswrapper[5018]: I1014 07:01:41.887822 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerDied","Data":"5799aa1d68f43f38935f4ea8e2f0276c5b71e0a9fe4eae041df29ae95f675905"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.616324 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99edde9a-4116-4bdf-84ff-863e9e6ed55e" path="/var/lib/kubelet/pods/99edde9a-4116-4bdf-84ff-863e9e6ed55e/volumes" Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899749 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"6903d0c7a7b1c5b05b7d5d92a7b9b7e7b51fdc3e0a41a004d691052a9fb01c87"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899807 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"7f6f0d708f967f55a79871cc26f3d6215d5cacd983c39a70e664f9bab7a83ec1"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899828 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"a037725b8c0a50c3c207571d8b05bc32711a72ad23873bb4f36e18df6ce18d6e"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899847 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"fdd36c41da5ea1b146947bd1680f175a46a4420392a5ab91f4e5993ed6ff676f"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899864 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"2c193b0b385e55cc47698e23eb64e4140f898857a4db5785bebdcdb0be6ce541"} Oct 14 07:01:42 crc kubenswrapper[5018]: I1014 07:01:42.899882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"3a434577e8512e9e319ff627a52393f5d8d552c00e1096d3f3ac2ba1d58bddc3"} Oct 14 07:01:44 crc kubenswrapper[5018]: I1014 07:01:44.915339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"357167cc55ee1738a5fdc30de972e50e52c4374c01ae19eaf9933f84547cd3e8"} Oct 14 07:01:47 crc kubenswrapper[5018]: I1014 07:01:47.936676 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" event={"ID":"847d2023-88e3-4116-af02-621ce12380a2","Type":"ContainerStarted","Data":"c0658ffd756521998de5b8b8e111f376ff1d4f96cc7aeca00e8972bce058293a"} Oct 14 07:01:47 crc kubenswrapper[5018]: I1014 07:01:47.937259 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:47 crc kubenswrapper[5018]: I1014 07:01:47.966171 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" podStartSLOduration=7.966152612 podStartE2EDuration="7.966152612s" podCreationTimestamp="2025-10-14 07:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:01:47.961713568 +0000 UTC m=+724.545760205" watchObservedRunningTime="2025-10-14 07:01:47.966152612 +0000 UTC m=+724.550199239" Oct 14 07:01:47 crc kubenswrapper[5018]: I1014 07:01:47.972811 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:48 crc kubenswrapper[5018]: I1014 07:01:48.942189 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:48 crc kubenswrapper[5018]: I1014 07:01:48.942511 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:48 crc kubenswrapper[5018]: I1014 07:01:48.974156 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:01:51 crc kubenswrapper[5018]: I1014 07:01:51.605675 5018 scope.go:117] "RemoveContainer" containerID="0eb191c53d3c98d15cfe80fae0123a7b7d712dc828106085b6059d32258a3d98" Oct 14 07:01:51 crc kubenswrapper[5018]: I1014 07:01:51.961486 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r7ftx_f1aa622b-96ac-42aa-8bd1-8fa174e65602/kube-multus/2.log" Oct 14 07:01:51 crc kubenswrapper[5018]: I1014 07:01:51.961806 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r7ftx" event={"ID":"f1aa622b-96ac-42aa-8bd1-8fa174e65602","Type":"ContainerStarted","Data":"ea2c08d90aa36f6c7045ee754ef6b71e7af062a2e61e90849153e7252d93fd37"} Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.547642 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-dtwvt"] Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.548588 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.551934 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.552187 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.553116 5018 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5sfhc" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.553492 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.569010 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dtwvt"] Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.689120 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mwjt\" (UniqueName: \"kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.689214 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.689514 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.790844 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.790962 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mwjt\" (UniqueName: \"kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.791002 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.791266 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.792770 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.828317 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mwjt\" (UniqueName: \"kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt\") pod \"crc-storage-crc-dtwvt\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.870099 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.894411 5018 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(b8adfcc3f98be93d8b01404d4243d01cb83339bb75aa62be051787efb5239aa5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.894766 5018 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(b8adfcc3f98be93d8b01404d4243d01cb83339bb75aa62be051787efb5239aa5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.894941 5018 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(b8adfcc3f98be93d8b01404d4243d01cb83339bb75aa62be051787efb5239aa5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.895134 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-dtwvt_crc-storage(9f0ac94c-3256-4e42-9520-a7af1cc2266e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-dtwvt_crc-storage(9f0ac94c-3256-4e42-9520-a7af1cc2266e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(b8adfcc3f98be93d8b01404d4243d01cb83339bb75aa62be051787efb5239aa5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-dtwvt" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.966493 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: I1014 07:01:52.966946 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.990861 5018 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(7d01f30eb1e24ca999625b579518e6aab26442f93276f1d5a03adbd103ad88b5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.990902 5018 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(7d01f30eb1e24ca999625b579518e6aab26442f93276f1d5a03adbd103ad88b5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.990920 5018 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(7d01f30eb1e24ca999625b579518e6aab26442f93276f1d5a03adbd103ad88b5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:01:52 crc kubenswrapper[5018]: E1014 07:01:52.990955 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-dtwvt_crc-storage(9f0ac94c-3256-4e42-9520-a7af1cc2266e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-dtwvt_crc-storage(9f0ac94c-3256-4e42-9520-a7af1cc2266e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dtwvt_crc-storage_9f0ac94c-3256-4e42-9520-a7af1cc2266e_0(7d01f30eb1e24ca999625b579518e6aab26442f93276f1d5a03adbd103ad88b5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-dtwvt" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" Oct 14 07:02:02 crc kubenswrapper[5018]: I1014 07:02:02.464073 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:02:02 crc kubenswrapper[5018]: I1014 07:02:02.464816 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:02:04 crc kubenswrapper[5018]: I1014 07:02:04.604581 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:02:04 crc kubenswrapper[5018]: I1014 07:02:04.609517 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:02:04 crc kubenswrapper[5018]: I1014 07:02:04.856704 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dtwvt"] Oct 14 07:02:04 crc kubenswrapper[5018]: W1014 07:02:04.864717 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f0ac94c_3256_4e42_9520_a7af1cc2266e.slice/crio-faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497 WatchSource:0}: Error finding container faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497: Status 404 returned error can't find the container with id faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497 Oct 14 07:02:04 crc kubenswrapper[5018]: I1014 07:02:04.866764 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:02:05 crc kubenswrapper[5018]: I1014 07:02:05.039881 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dtwvt" event={"ID":"9f0ac94c-3256-4e42-9520-a7af1cc2266e","Type":"ContainerStarted","Data":"faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497"} Oct 14 07:02:07 crc kubenswrapper[5018]: I1014 07:02:07.053799 5018 generic.go:334] "Generic (PLEG): container finished" podID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" containerID="65eb8594532e20620d327b9aea09bf441323d01f58be10e38d16566bc934e879" exitCode=0 Oct 14 07:02:07 crc kubenswrapper[5018]: I1014 07:02:07.053905 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dtwvt" event={"ID":"9f0ac94c-3256-4e42-9520-a7af1cc2266e","Type":"ContainerDied","Data":"65eb8594532e20620d327b9aea09bf441323d01f58be10e38d16566bc934e879"} Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.299792 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.402756 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage\") pod \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.402793 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt\") pod \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.402859 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mwjt\" (UniqueName: \"kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt\") pod \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\" (UID: \"9f0ac94c-3256-4e42-9520-a7af1cc2266e\") " Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.402968 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9f0ac94c-3256-4e42-9520-a7af1cc2266e" (UID: "9f0ac94c-3256-4e42-9520-a7af1cc2266e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.403193 5018 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9f0ac94c-3256-4e42-9520-a7af1cc2266e-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.409834 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt" (OuterVolumeSpecName: "kube-api-access-5mwjt") pod "9f0ac94c-3256-4e42-9520-a7af1cc2266e" (UID: "9f0ac94c-3256-4e42-9520-a7af1cc2266e"). InnerVolumeSpecName "kube-api-access-5mwjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.427762 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9f0ac94c-3256-4e42-9520-a7af1cc2266e" (UID: "9f0ac94c-3256-4e42-9520-a7af1cc2266e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.504817 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mwjt\" (UniqueName: \"kubernetes.io/projected/9f0ac94c-3256-4e42-9520-a7af1cc2266e-kube-api-access-5mwjt\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:08 crc kubenswrapper[5018]: I1014 07:02:08.504852 5018 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9f0ac94c-3256-4e42-9520-a7af1cc2266e-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.068607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dtwvt" event={"ID":"9f0ac94c-3256-4e42-9520-a7af1cc2266e","Type":"ContainerDied","Data":"faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497"} Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.068671 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dtwvt" Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.068685 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faa3c971dc12f62467f71bab493a6f834292bad35127ca832ffe6a14f42dc497" Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.555900 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.556128 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" containerID="cri-o://524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930" gracePeriod=30 Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.662105 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.662874 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" podUID="17655b20-7a11-4fa4-a438-994408b65861" containerName="route-controller-manager" containerID="cri-o://86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c" gracePeriod=30 Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.956584 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 07:02:09 crc kubenswrapper[5018]: I1014 07:02:09.986329 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.072902 5018 generic.go:334] "Generic (PLEG): container finished" podID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerID="524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930" exitCode=0 Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.072953 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" event={"ID":"ac8f3d02-ac5f-4879-a8d3-66c226c42964","Type":"ContainerDied","Data":"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930"} Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.072977 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" event={"ID":"ac8f3d02-ac5f-4879-a8d3-66c226c42964","Type":"ContainerDied","Data":"445fc5522c9dfdc02307597880bfac99d795c74ac8a92db8d6a71819f843d6ab"} Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.072994 5018 scope.go:117] "RemoveContainer" containerID="524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.073077 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6vg2h" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.075866 5018 generic.go:334] "Generic (PLEG): container finished" podID="17655b20-7a11-4fa4-a438-994408b65861" containerID="86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c" exitCode=0 Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.075912 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" event={"ID":"17655b20-7a11-4fa4-a438-994408b65861","Type":"ContainerDied","Data":"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c"} Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.075954 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" event={"ID":"17655b20-7a11-4fa4-a438-994408b65861","Type":"ContainerDied","Data":"f0df6eef91a6ee36d93ef7154656b17ec7f1573506910a34acd3cceabcdcb4fd"} Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.075961 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.090260 5018 scope.go:117] "RemoveContainer" containerID="524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930" Oct 14 07:02:10 crc kubenswrapper[5018]: E1014 07:02:10.090799 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930\": container with ID starting with 524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930 not found: ID does not exist" containerID="524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.090837 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930"} err="failed to get container status \"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930\": rpc error: code = NotFound desc = could not find container \"524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930\": container with ID starting with 524396142f3096a731dc035fb05bac6d6c57361b0d4c5bc9e0a06cddb84a7930 not found: ID does not exist" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.090864 5018 scope.go:117] "RemoveContainer" containerID="86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.110383 5018 scope.go:117] "RemoveContainer" containerID="86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c" Oct 14 07:02:10 crc kubenswrapper[5018]: E1014 07:02:10.110850 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c\": container with ID starting with 86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c not found: ID does not exist" containerID="86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.110875 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c"} err="failed to get container status \"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c\": rpc error: code = NotFound desc = could not find container \"86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c\": container with ID starting with 86aee3a57e886c0114335f4806014af94f4d090f1100a34d9580a8e66e97473c not found: ID does not exist" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126529 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config\") pod \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126571 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca\") pod \"17655b20-7a11-4fa4-a438-994408b65861\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126606 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles\") pod \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126641 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czt2r\" (UniqueName: \"kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r\") pod \"17655b20-7a11-4fa4-a438-994408b65861\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126675 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttmf\" (UniqueName: \"kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf\") pod \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126696 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca\") pod \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126745 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert\") pod \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\" (UID: \"ac8f3d02-ac5f-4879-a8d3-66c226c42964\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126765 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config\") pod \"17655b20-7a11-4fa4-a438-994408b65861\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.126791 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert\") pod \"17655b20-7a11-4fa4-a438-994408b65861\" (UID: \"17655b20-7a11-4fa4-a438-994408b65861\") " Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.127791 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca" (OuterVolumeSpecName: "client-ca") pod "ac8f3d02-ac5f-4879-a8d3-66c226c42964" (UID: "ac8f3d02-ac5f-4879-a8d3-66c226c42964"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.127809 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca" (OuterVolumeSpecName: "client-ca") pod "17655b20-7a11-4fa4-a438-994408b65861" (UID: "17655b20-7a11-4fa4-a438-994408b65861"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.127916 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ac8f3d02-ac5f-4879-a8d3-66c226c42964" (UID: "ac8f3d02-ac5f-4879-a8d3-66c226c42964"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.127995 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config" (OuterVolumeSpecName: "config") pod "17655b20-7a11-4fa4-a438-994408b65861" (UID: "17655b20-7a11-4fa4-a438-994408b65861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.128260 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config" (OuterVolumeSpecName: "config") pod "ac8f3d02-ac5f-4879-a8d3-66c226c42964" (UID: "ac8f3d02-ac5f-4879-a8d3-66c226c42964"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.131511 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "17655b20-7a11-4fa4-a438-994408b65861" (UID: "17655b20-7a11-4fa4-a438-994408b65861"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.131602 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r" (OuterVolumeSpecName: "kube-api-access-czt2r") pod "17655b20-7a11-4fa4-a438-994408b65861" (UID: "17655b20-7a11-4fa4-a438-994408b65861"). InnerVolumeSpecName "kube-api-access-czt2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.132474 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf" (OuterVolumeSpecName: "kube-api-access-7ttmf") pod "ac8f3d02-ac5f-4879-a8d3-66c226c42964" (UID: "ac8f3d02-ac5f-4879-a8d3-66c226c42964"). InnerVolumeSpecName "kube-api-access-7ttmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.132877 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ac8f3d02-ac5f-4879-a8d3-66c226c42964" (UID: "ac8f3d02-ac5f-4879-a8d3-66c226c42964"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.228860 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.229312 5018 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.229671 5018 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.229953 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czt2r\" (UniqueName: \"kubernetes.io/projected/17655b20-7a11-4fa4-a438-994408b65861-kube-api-access-czt2r\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.230141 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttmf\" (UniqueName: \"kubernetes.io/projected/ac8f3d02-ac5f-4879-a8d3-66c226c42964-kube-api-access-7ttmf\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.230314 5018 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8f3d02-ac5f-4879-a8d3-66c226c42964-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.231461 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8f3d02-ac5f-4879-a8d3-66c226c42964-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.232315 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17655b20-7a11-4fa4-a438-994408b65861-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.232484 5018 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17655b20-7a11-4fa4-a438-994408b65861-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.425978 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.428454 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6vg2h"] Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.449137 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.454291 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zq8rz"] Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.619597 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17655b20-7a11-4fa4-a438-994408b65861" path="/var/lib/kubelet/pods/17655b20-7a11-4fa4-a438-994408b65861/volumes" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.621151 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" path="/var/lib/kubelet/pods/ac8f3d02-ac5f-4879-a8d3-66c226c42964/volumes" Oct 14 07:02:10 crc kubenswrapper[5018]: I1014 07:02:10.812181 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ddbv7" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.090789 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s"] Oct 14 07:02:11 crc kubenswrapper[5018]: E1014 07:02:11.091162 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17655b20-7a11-4fa4-a438-994408b65861" containerName="route-controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091189 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17655b20-7a11-4fa4-a438-994408b65861" containerName="route-controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: E1014 07:02:11.091229 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" containerName="storage" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091242 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" containerName="storage" Oct 14 07:02:11 crc kubenswrapper[5018]: E1014 07:02:11.091267 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091279 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091429 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" containerName="storage" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091457 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="17655b20-7a11-4fa4-a438-994408b65861" containerName="route-controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.091474 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8f3d02-ac5f-4879-a8d3-66c226c42964" containerName="controller-manager" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.092059 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.093838 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.093970 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc"] Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.094523 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.094591 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.094812 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.095321 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.095416 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.096022 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.096759 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.097207 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.097453 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.097469 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.098237 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.098256 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.105689 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.106689 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s"] Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.108913 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc"] Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243099 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-client-ca\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243145 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-proxy-ca-bundles\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243168 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3e5bb36-cf55-492d-b945-c7e00419358c-serving-cert\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243184 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcdks\" (UniqueName: \"kubernetes.io/projected/a3e5bb36-cf55-492d-b945-c7e00419358c-kube-api-access-vcdks\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243210 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e0793a-93b9-4cac-b057-0d3da030f5e2-serving-cert\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243229 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9r2k\" (UniqueName: \"kubernetes.io/projected/b3e0793a-93b9-4cac-b057-0d3da030f5e2-kube-api-access-s9r2k\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243366 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-config\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243427 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-client-ca\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.243600 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-config\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345192 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e0793a-93b9-4cac-b057-0d3da030f5e2-serving-cert\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345234 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9r2k\" (UniqueName: \"kubernetes.io/projected/b3e0793a-93b9-4cac-b057-0d3da030f5e2-kube-api-access-s9r2k\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345267 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-config\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345284 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-client-ca\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345340 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-config\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345369 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-client-ca\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345395 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-proxy-ca-bundles\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345414 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3e5bb36-cf55-492d-b945-c7e00419358c-serving-cert\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.345435 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcdks\" (UniqueName: \"kubernetes.io/projected/a3e5bb36-cf55-492d-b945-c7e00419358c-kube-api-access-vcdks\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.346429 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-proxy-ca-bundles\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.346486 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-client-ca\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.346507 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-client-ca\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.346670 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e5bb36-cf55-492d-b945-c7e00419358c-config\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.347293 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3e0793a-93b9-4cac-b057-0d3da030f5e2-config\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.349269 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3e0793a-93b9-4cac-b057-0d3da030f5e2-serving-cert\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.349282 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3e5bb36-cf55-492d-b945-c7e00419358c-serving-cert\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.363703 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9r2k\" (UniqueName: \"kubernetes.io/projected/b3e0793a-93b9-4cac-b057-0d3da030f5e2-kube-api-access-s9r2k\") pod \"route-controller-manager-548c95b55c-hv9pc\" (UID: \"b3e0793a-93b9-4cac-b057-0d3da030f5e2\") " pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.366997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcdks\" (UniqueName: \"kubernetes.io/projected/a3e5bb36-cf55-492d-b945-c7e00419358c-kube-api-access-vcdks\") pod \"controller-manager-7c74fd4f9b-x8z8s\" (UID: \"a3e5bb36-cf55-492d-b945-c7e00419358c\") " pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.428424 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.441187 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.624571 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s"] Oct 14 07:02:11 crc kubenswrapper[5018]: I1014 07:02:11.656854 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc"] Oct 14 07:02:11 crc kubenswrapper[5018]: W1014 07:02:11.658740 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3e0793a_93b9_4cac_b057_0d3da030f5e2.slice/crio-1cbdb277574c516a9de2462eb23e9f95610f531c96b8dcee7a39733b60d91aaf WatchSource:0}: Error finding container 1cbdb277574c516a9de2462eb23e9f95610f531c96b8dcee7a39733b60d91aaf: Status 404 returned error can't find the container with id 1cbdb277574c516a9de2462eb23e9f95610f531c96b8dcee7a39733b60d91aaf Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.090464 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" event={"ID":"b3e0793a-93b9-4cac-b057-0d3da030f5e2","Type":"ContainerStarted","Data":"12574c0eeb41c30f69ba01aba84dec7f3c6ce55ad22d1f09b11e794a25b53678"} Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.090841 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" event={"ID":"b3e0793a-93b9-4cac-b057-0d3da030f5e2","Type":"ContainerStarted","Data":"1cbdb277574c516a9de2462eb23e9f95610f531c96b8dcee7a39733b60d91aaf"} Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.090862 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.091891 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" event={"ID":"a3e5bb36-cf55-492d-b945-c7e00419358c","Type":"ContainerStarted","Data":"45392077e95e796e202a1343ffc44bdd72d9aa617a83a16dc3379f19eaf5a6c5"} Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.091921 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" event={"ID":"a3e5bb36-cf55-492d-b945-c7e00419358c","Type":"ContainerStarted","Data":"56bc0e0632e8b7dc1c738f33cf4e5efc0704674b9bb98c88b4259726e8bab595"} Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.092355 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.097528 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.115012 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" podStartSLOduration=3.114993695 podStartE2EDuration="3.114993695s" podCreationTimestamp="2025-10-14 07:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:12.113823572 +0000 UTC m=+748.697870209" watchObservedRunningTime="2025-10-14 07:02:12.114993695 +0000 UTC m=+748.699040332" Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.131577 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c74fd4f9b-x8z8s" podStartSLOduration=3.131561078 podStartE2EDuration="3.131561078s" podCreationTimestamp="2025-10-14 07:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:12.131082134 +0000 UTC m=+748.715128771" watchObservedRunningTime="2025-10-14 07:02:12.131561078 +0000 UTC m=+748.715607715" Oct 14 07:02:12 crc kubenswrapper[5018]: I1014 07:02:12.746603 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-548c95b55c-hv9pc" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.739561 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd"] Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.741091 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.743879 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.752081 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd"] Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.815497 5018 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.923823 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.923938 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:16 crc kubenswrapper[5018]: I1014 07:02:16.923997 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwhkj\" (UniqueName: \"kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.025545 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.025934 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwhkj\" (UniqueName: \"kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.026078 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.026189 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.026497 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.044333 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwhkj\" (UniqueName: \"kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.055830 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:17 crc kubenswrapper[5018]: I1014 07:02:17.480430 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd"] Oct 14 07:02:17 crc kubenswrapper[5018]: W1014 07:02:17.487610 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf742418e_85fe_4763_9fc5_09795c16569b.slice/crio-7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c WatchSource:0}: Error finding container 7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c: Status 404 returned error can't find the container with id 7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c Oct 14 07:02:18 crc kubenswrapper[5018]: I1014 07:02:18.137095 5018 generic.go:334] "Generic (PLEG): container finished" podID="f742418e-85fe-4763-9fc5-09795c16569b" containerID="3f48fe95cd1612eb523fa81c88228c5435d5b7445195c2c72cfaad1804850521" exitCode=0 Oct 14 07:02:18 crc kubenswrapper[5018]: I1014 07:02:18.137185 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" event={"ID":"f742418e-85fe-4763-9fc5-09795c16569b","Type":"ContainerDied","Data":"3f48fe95cd1612eb523fa81c88228c5435d5b7445195c2c72cfaad1804850521"} Oct 14 07:02:18 crc kubenswrapper[5018]: I1014 07:02:18.137420 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" event={"ID":"f742418e-85fe-4763-9fc5-09795c16569b","Type":"ContainerStarted","Data":"7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c"} Oct 14 07:02:18 crc kubenswrapper[5018]: I1014 07:02:18.994972 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:18 crc kubenswrapper[5018]: I1014 07:02:18.997826 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.012866 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.161198 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.161301 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.161374 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk7hg\" (UniqueName: \"kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.262918 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk7hg\" (UniqueName: \"kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.263025 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.263077 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.263640 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.263776 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.289383 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk7hg\" (UniqueName: \"kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg\") pod \"redhat-operators-cvt4g\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.324557 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:19 crc kubenswrapper[5018]: I1014 07:02:19.820321 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:19 crc kubenswrapper[5018]: W1014 07:02:19.827734 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74203cf3_9c19_4bd3_9d1b_040a99893ab1.slice/crio-d131d9e6632780f6617d3c79c311eb78fa03ca587fd78fb79369e9fc8213045b WatchSource:0}: Error finding container d131d9e6632780f6617d3c79c311eb78fa03ca587fd78fb79369e9fc8213045b: Status 404 returned error can't find the container with id d131d9e6632780f6617d3c79c311eb78fa03ca587fd78fb79369e9fc8213045b Oct 14 07:02:20 crc kubenswrapper[5018]: I1014 07:02:20.150885 5018 generic.go:334] "Generic (PLEG): container finished" podID="f742418e-85fe-4763-9fc5-09795c16569b" containerID="79d87d90701336742a8df11e9aa28c3452653a689cb02395841b56561e45a313" exitCode=0 Oct 14 07:02:20 crc kubenswrapper[5018]: I1014 07:02:20.150934 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" event={"ID":"f742418e-85fe-4763-9fc5-09795c16569b","Type":"ContainerDied","Data":"79d87d90701336742a8df11e9aa28c3452653a689cb02395841b56561e45a313"} Oct 14 07:02:20 crc kubenswrapper[5018]: I1014 07:02:20.154732 5018 generic.go:334] "Generic (PLEG): container finished" podID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerID="b1bdff870eaa3fca906238bccdeca4044ceab671e0aa358171497acb81201857" exitCode=0 Oct 14 07:02:20 crc kubenswrapper[5018]: I1014 07:02:20.154770 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerDied","Data":"b1bdff870eaa3fca906238bccdeca4044ceab671e0aa358171497acb81201857"} Oct 14 07:02:20 crc kubenswrapper[5018]: I1014 07:02:20.154794 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerStarted","Data":"d131d9e6632780f6617d3c79c311eb78fa03ca587fd78fb79369e9fc8213045b"} Oct 14 07:02:21 crc kubenswrapper[5018]: I1014 07:02:21.166287 5018 generic.go:334] "Generic (PLEG): container finished" podID="f742418e-85fe-4763-9fc5-09795c16569b" containerID="cf7eade4b41e9e2a173f48ce581c68628c9a4004af8778911bd9150eb5a9af4f" exitCode=0 Oct 14 07:02:21 crc kubenswrapper[5018]: I1014 07:02:21.166349 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" event={"ID":"f742418e-85fe-4763-9fc5-09795c16569b","Type":"ContainerDied","Data":"cf7eade4b41e9e2a173f48ce581c68628c9a4004af8778911bd9150eb5a9af4f"} Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.180108 5018 generic.go:334] "Generic (PLEG): container finished" podID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerID="870aa281926e5f30fef860758d82c17a71cd7f846b8073c0bbb2803332f99cff" exitCode=0 Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.180238 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerDied","Data":"870aa281926e5f30fef860758d82c17a71cd7f846b8073c0bbb2803332f99cff"} Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.606048 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.709861 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwhkj\" (UniqueName: \"kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj\") pod \"f742418e-85fe-4763-9fc5-09795c16569b\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.709938 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util\") pod \"f742418e-85fe-4763-9fc5-09795c16569b\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.709996 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle\") pod \"f742418e-85fe-4763-9fc5-09795c16569b\" (UID: \"f742418e-85fe-4763-9fc5-09795c16569b\") " Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.710855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle" (OuterVolumeSpecName: "bundle") pod "f742418e-85fe-4763-9fc5-09795c16569b" (UID: "f742418e-85fe-4763-9fc5-09795c16569b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.716631 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj" (OuterVolumeSpecName: "kube-api-access-lwhkj") pod "f742418e-85fe-4763-9fc5-09795c16569b" (UID: "f742418e-85fe-4763-9fc5-09795c16569b"). InnerVolumeSpecName "kube-api-access-lwhkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.743831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util" (OuterVolumeSpecName: "util") pod "f742418e-85fe-4763-9fc5-09795c16569b" (UID: "f742418e-85fe-4763-9fc5-09795c16569b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.811529 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwhkj\" (UniqueName: \"kubernetes.io/projected/f742418e-85fe-4763-9fc5-09795c16569b-kube-api-access-lwhkj\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.812022 5018 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:22 crc kubenswrapper[5018]: I1014 07:02:22.812074 5018 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f742418e-85fe-4763-9fc5-09795c16569b-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:23 crc kubenswrapper[5018]: I1014 07:02:23.192447 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" event={"ID":"f742418e-85fe-4763-9fc5-09795c16569b","Type":"ContainerDied","Data":"7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c"} Oct 14 07:02:23 crc kubenswrapper[5018]: I1014 07:02:23.192488 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cf0616b275d3365daa66aaa17a256eca3bed47701559230841c4ed5a06e154c" Oct 14 07:02:23 crc kubenswrapper[5018]: I1014 07:02:23.192486 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd" Oct 14 07:02:23 crc kubenswrapper[5018]: I1014 07:02:23.195597 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerStarted","Data":"52dcaa037cf33580f3d267e9a6c30d361bf74086c2ef26b99fdae66386a3e70f"} Oct 14 07:02:23 crc kubenswrapper[5018]: I1014 07:02:23.224439 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cvt4g" podStartSLOduration=2.736278072 podStartE2EDuration="5.224413943s" podCreationTimestamp="2025-10-14 07:02:18 +0000 UTC" firstStartedPulling="2025-10-14 07:02:20.156127608 +0000 UTC m=+756.740174235" lastFinishedPulling="2025-10-14 07:02:22.644263479 +0000 UTC m=+759.228310106" observedRunningTime="2025-10-14 07:02:23.224240718 +0000 UTC m=+759.808287355" watchObservedRunningTime="2025-10-14 07:02:23.224413943 +0000 UTC m=+759.808460610" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.911969 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk"] Oct 14 07:02:24 crc kubenswrapper[5018]: E1014 07:02:24.912558 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="extract" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.912572 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="extract" Oct 14 07:02:24 crc kubenswrapper[5018]: E1014 07:02:24.912585 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="util" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.912592 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="util" Oct 14 07:02:24 crc kubenswrapper[5018]: E1014 07:02:24.912603 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="pull" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.912612 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="pull" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.912749 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f742418e-85fe-4763-9fc5-09795c16569b" containerName="extract" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.913162 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.914751 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.915229 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mbq64" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.916158 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 14 07:02:24 crc kubenswrapper[5018]: I1014 07:02:24.936992 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk"] Oct 14 07:02:25 crc kubenswrapper[5018]: I1014 07:02:25.041498 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfwsn\" (UniqueName: \"kubernetes.io/projected/b06cab00-8e52-4c9a-b263-e54c98b12aa5-kube-api-access-nfwsn\") pod \"nmstate-operator-858ddd8f98-8j4wk\" (UID: \"b06cab00-8e52-4c9a-b263-e54c98b12aa5\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" Oct 14 07:02:25 crc kubenswrapper[5018]: I1014 07:02:25.143250 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfwsn\" (UniqueName: \"kubernetes.io/projected/b06cab00-8e52-4c9a-b263-e54c98b12aa5-kube-api-access-nfwsn\") pod \"nmstate-operator-858ddd8f98-8j4wk\" (UID: \"b06cab00-8e52-4c9a-b263-e54c98b12aa5\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" Oct 14 07:02:25 crc kubenswrapper[5018]: I1014 07:02:25.174494 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfwsn\" (UniqueName: \"kubernetes.io/projected/b06cab00-8e52-4c9a-b263-e54c98b12aa5-kube-api-access-nfwsn\") pod \"nmstate-operator-858ddd8f98-8j4wk\" (UID: \"b06cab00-8e52-4c9a-b263-e54c98b12aa5\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" Oct 14 07:02:25 crc kubenswrapper[5018]: I1014 07:02:25.229838 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" Oct 14 07:02:25 crc kubenswrapper[5018]: I1014 07:02:25.642368 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk"] Oct 14 07:02:26 crc kubenswrapper[5018]: I1014 07:02:26.218357 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" event={"ID":"b06cab00-8e52-4c9a-b263-e54c98b12aa5","Type":"ContainerStarted","Data":"840e45b227db396749b7efccb20d6b57a2f652cc587df607e6f2ab0c39f82a4a"} Oct 14 07:02:29 crc kubenswrapper[5018]: I1014 07:02:29.241469 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" event={"ID":"b06cab00-8e52-4c9a-b263-e54c98b12aa5","Type":"ContainerStarted","Data":"77ab723b8b5509131fbfd8df5a57353a5be621da3efdb05966ad07e233fb3def"} Oct 14 07:02:29 crc kubenswrapper[5018]: I1014 07:02:29.277387 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-8j4wk" podStartSLOduration=2.8123307520000003 podStartE2EDuration="5.277361138s" podCreationTimestamp="2025-10-14 07:02:24 +0000 UTC" firstStartedPulling="2025-10-14 07:02:25.657201469 +0000 UTC m=+762.241248116" lastFinishedPulling="2025-10-14 07:02:28.122231865 +0000 UTC m=+764.706278502" observedRunningTime="2025-10-14 07:02:29.26453832 +0000 UTC m=+765.848584957" watchObservedRunningTime="2025-10-14 07:02:29.277361138 +0000 UTC m=+765.861407805" Oct 14 07:02:29 crc kubenswrapper[5018]: I1014 07:02:29.325210 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:29 crc kubenswrapper[5018]: I1014 07:02:29.325304 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:29 crc kubenswrapper[5018]: I1014 07:02:29.388106 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.229733 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.231184 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.235993 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-mh7cp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.241797 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.250488 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.253261 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.257104 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.263776 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.284210 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-556sb"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.284897 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.309873 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bck6h\" (UniqueName: \"kubernetes.io/projected/c718fdd6-a84c-4842-aa5d-f6ccc0c8babd-kube-api-access-bck6h\") pod \"nmstate-metrics-fdff9cb8d-lszjp\" (UID: \"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.321891 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.363087 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.363748 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.365181 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-5rglc" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.365458 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.365789 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.372235 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411164 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-dbus-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411447 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzv7z\" (UniqueName: \"kubernetes.io/projected/c32afd97-261a-4e41-bda2-ae5de91f5d6f-kube-api-access-hzv7z\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411536 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-ovs-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411667 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74782f46-6533-44c4-b5bf-bcdd43231af6-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411767 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-nmstate-lock\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411841 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bck6h\" (UniqueName: \"kubernetes.io/projected/c718fdd6-a84c-4842-aa5d-f6ccc0c8babd-kube-api-access-bck6h\") pod \"nmstate-metrics-fdff9cb8d-lszjp\" (UID: \"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.411935 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrt64\" (UniqueName: \"kubernetes.io/projected/74782f46-6533-44c4-b5bf-bcdd43231af6-kube-api-access-hrt64\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.431802 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bck6h\" (UniqueName: \"kubernetes.io/projected/c718fdd6-a84c-4842-aa5d-f6ccc0c8babd-kube-api-access-bck6h\") pod \"nmstate-metrics-fdff9cb8d-lszjp\" (UID: \"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.513236 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-dbus-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.513664 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-dbus-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.513343 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzv7z\" (UniqueName: \"kubernetes.io/projected/c32afd97-261a-4e41-bda2-ae5de91f5d6f-kube-api-access-hzv7z\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.513800 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-ovs-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.513865 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-ovs-socket\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.514009 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74782f46-6533-44c4-b5bf-bcdd43231af6-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.514044 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvqqm\" (UniqueName: \"kubernetes.io/projected/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-kube-api-access-dvqqm\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.514959 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-nmstate-lock\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.515021 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.515069 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c32afd97-261a-4e41-bda2-ae5de91f5d6f-nmstate-lock\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.515053 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrt64\" (UniqueName: \"kubernetes.io/projected/74782f46-6533-44c4-b5bf-bcdd43231af6-kube-api-access-hrt64\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.515227 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.518401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74782f46-6533-44c4-b5bf-bcdd43231af6-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.530055 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzv7z\" (UniqueName: \"kubernetes.io/projected/c32afd97-261a-4e41-bda2-ae5de91f5d6f-kube-api-access-hzv7z\") pod \"nmstate-handler-556sb\" (UID: \"c32afd97-261a-4e41-bda2-ae5de91f5d6f\") " pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.533355 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrt64\" (UniqueName: \"kubernetes.io/projected/74782f46-6533-44c4-b5bf-bcdd43231af6-kube-api-access-hrt64\") pod \"nmstate-webhook-6cdbc54649-rcf8x\" (UID: \"74782f46-6533-44c4-b5bf-bcdd43231af6\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.551906 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.555790 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-bddbd9cfb-t9fh7"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.557089 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.570784 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bddbd9cfb-t9fh7"] Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.584503 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.606509 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.616226 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvqqm\" (UniqueName: \"kubernetes.io/projected/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-kube-api-access-dvqqm\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.616276 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.616313 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.618535 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.624706 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.636495 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvqqm\" (UniqueName: \"kubernetes.io/projected/b6eb676d-62fd-4035-8cfd-4617d1cb7b35-kube-api-access-dvqqm\") pod \"nmstate-console-plugin-6b874cbd85-gbl7v\" (UID: \"b6eb676d-62fd-4035-8cfd-4617d1cb7b35\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: W1014 07:02:30.643476 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32afd97_261a_4e41_bda2_ae5de91f5d6f.slice/crio-0ba3b495ea2992a883449f42034d185e93d85bac88d5cb76a1046e9696aabbb3 WatchSource:0}: Error finding container 0ba3b495ea2992a883449f42034d185e93d85bac88d5cb76a1046e9696aabbb3: Status 404 returned error can't find the container with id 0ba3b495ea2992a883449f42034d185e93d85bac88d5cb76a1046e9696aabbb3 Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.682054 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717252 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-service-ca\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717302 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-console-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717354 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717382 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmc8t\" (UniqueName: \"kubernetes.io/projected/63909d46-66cf-4cbb-a8de-82078f064384-kube-api-access-xmc8t\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717422 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-trusted-ca-bundle\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717457 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-oauth-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.717486 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-oauth-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818845 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-service-ca\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818885 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-console-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818916 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818934 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmc8t\" (UniqueName: \"kubernetes.io/projected/63909d46-66cf-4cbb-a8de-82078f064384-kube-api-access-xmc8t\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818959 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-trusted-ca-bundle\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.818984 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-oauth-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.819000 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-oauth-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.848789 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-service-ca\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.851121 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-oauth-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.852339 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-console-config\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.852918 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-trusted-ca-bundle\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.853238 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/63909d46-66cf-4cbb-a8de-82078f064384-oauth-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.854855 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/63909d46-66cf-4cbb-a8de-82078f064384-console-serving-cert\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.856069 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmc8t\" (UniqueName: \"kubernetes.io/projected/63909d46-66cf-4cbb-a8de-82078f064384-kube-api-access-xmc8t\") pod \"console-bddbd9cfb-t9fh7\" (UID: \"63909d46-66cf-4cbb-a8de-82078f064384\") " pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.898336 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x"] Oct 14 07:02:30 crc kubenswrapper[5018]: W1014 07:02:30.901224 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74782f46_6533_44c4_b5bf_bcdd43231af6.slice/crio-88ab1ce42ce152d5d56d6820e8c0b95564d4d9c394fb877a3140b39ba57c8ed4 WatchSource:0}: Error finding container 88ab1ce42ce152d5d56d6820e8c0b95564d4d9c394fb877a3140b39ba57c8ed4: Status 404 returned error can't find the container with id 88ab1ce42ce152d5d56d6820e8c0b95564d4d9c394fb877a3140b39ba57c8ed4 Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.925238 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:30 crc kubenswrapper[5018]: I1014 07:02:30.989064 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp"] Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.128112 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v"] Oct 14 07:02:31 crc kubenswrapper[5018]: W1014 07:02:31.133354 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6eb676d_62fd_4035_8cfd_4617d1cb7b35.slice/crio-01e08103353098c1e33e6b89431e1cc40315c2f35114891765848e66ce5a0481 WatchSource:0}: Error finding container 01e08103353098c1e33e6b89431e1cc40315c2f35114891765848e66ce5a0481: Status 404 returned error can't find the container with id 01e08103353098c1e33e6b89431e1cc40315c2f35114891765848e66ce5a0481 Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.184955 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.266245 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-556sb" event={"ID":"c32afd97-261a-4e41-bda2-ae5de91f5d6f","Type":"ContainerStarted","Data":"0ba3b495ea2992a883449f42034d185e93d85bac88d5cb76a1046e9696aabbb3"} Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.267106 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" event={"ID":"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd","Type":"ContainerStarted","Data":"1510004f586e9f9154102535dea981b6837620d7bee81727a5218ebf0cd706ed"} Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.268419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" event={"ID":"74782f46-6533-44c4-b5bf-bcdd43231af6","Type":"ContainerStarted","Data":"88ab1ce42ce152d5d56d6820e8c0b95564d4d9c394fb877a3140b39ba57c8ed4"} Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.272394 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" event={"ID":"b6eb676d-62fd-4035-8cfd-4617d1cb7b35","Type":"ContainerStarted","Data":"01e08103353098c1e33e6b89431e1cc40315c2f35114891765848e66ce5a0481"} Oct 14 07:02:31 crc kubenswrapper[5018]: I1014 07:02:31.307608 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bddbd9cfb-t9fh7"] Oct 14 07:02:31 crc kubenswrapper[5018]: W1014 07:02:31.312813 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63909d46_66cf_4cbb_a8de_82078f064384.slice/crio-4b1ccf98077c05b627d0186641d9485e21109c57671f39f49b5385b7b0c34ef3 WatchSource:0}: Error finding container 4b1ccf98077c05b627d0186641d9485e21109c57671f39f49b5385b7b0c34ef3: Status 404 returned error can't find the container with id 4b1ccf98077c05b627d0186641d9485e21109c57671f39f49b5385b7b0c34ef3 Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.283558 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cvt4g" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="registry-server" containerID="cri-o://52dcaa037cf33580f3d267e9a6c30d361bf74086c2ef26b99fdae66386a3e70f" gracePeriod=2 Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.283910 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bddbd9cfb-t9fh7" event={"ID":"63909d46-66cf-4cbb-a8de-82078f064384","Type":"ContainerStarted","Data":"c6027ab1d4eee605332d008e82056b14ed455cbbf0fcc019e51dd20c7d086a0b"} Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.283931 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bddbd9cfb-t9fh7" event={"ID":"63909d46-66cf-4cbb-a8de-82078f064384","Type":"ContainerStarted","Data":"4b1ccf98077c05b627d0186641d9485e21109c57671f39f49b5385b7b0c34ef3"} Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.303170 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-bddbd9cfb-t9fh7" podStartSLOduration=2.303150466 podStartE2EDuration="2.303150466s" podCreationTimestamp="2025-10-14 07:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:32.300581234 +0000 UTC m=+768.884627861" watchObservedRunningTime="2025-10-14 07:02:32.303150466 +0000 UTC m=+768.887197093" Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.463265 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.463343 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.463385 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.463919 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:02:32 crc kubenswrapper[5018]: I1014 07:02:32.463987 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345" gracePeriod=600 Oct 14 07:02:33 crc kubenswrapper[5018]: I1014 07:02:33.293528 5018 generic.go:334] "Generic (PLEG): container finished" podID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerID="52dcaa037cf33580f3d267e9a6c30d361bf74086c2ef26b99fdae66386a3e70f" exitCode=0 Oct 14 07:02:33 crc kubenswrapper[5018]: I1014 07:02:33.293587 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerDied","Data":"52dcaa037cf33580f3d267e9a6c30d361bf74086c2ef26b99fdae66386a3e70f"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.064765 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.162670 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content\") pod \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.163047 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities\") pod \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.163088 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk7hg\" (UniqueName: \"kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg\") pod \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\" (UID: \"74203cf3-9c19-4bd3-9d1b-040a99893ab1\") " Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.169472 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities" (OuterVolumeSpecName: "utilities") pod "74203cf3-9c19-4bd3-9d1b-040a99893ab1" (UID: "74203cf3-9c19-4bd3-9d1b-040a99893ab1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.170821 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg" (OuterVolumeSpecName: "kube-api-access-pk7hg") pod "74203cf3-9c19-4bd3-9d1b-040a99893ab1" (UID: "74203cf3-9c19-4bd3-9d1b-040a99893ab1"). InnerVolumeSpecName "kube-api-access-pk7hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.243881 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74203cf3-9c19-4bd3-9d1b-040a99893ab1" (UID: "74203cf3-9c19-4bd3-9d1b-040a99893ab1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.264174 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.264202 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74203cf3-9c19-4bd3-9d1b-040a99893ab1-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.264211 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk7hg\" (UniqueName: \"kubernetes.io/projected/74203cf3-9c19-4bd3-9d1b-040a99893ab1-kube-api-access-pk7hg\") on node \"crc\" DevicePath \"\"" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.306690 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" event={"ID":"74782f46-6533-44c4-b5bf-bcdd43231af6","Type":"ContainerStarted","Data":"4c19c533231501bd5815360a8c5982f8328d0d2fe8f326f4b1432c0bbe5b461c"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.307545 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.310170 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" event={"ID":"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd","Type":"ContainerStarted","Data":"c91e33042a46126fdabee70a5c2b2449492ed518394e965dd982c53dfb73dbf3"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.313308 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345" exitCode=0 Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.313347 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.313364 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.313380 5018 scope.go:117] "RemoveContainer" containerID="f38380272cb0e3168fa17670bb37726b64f8e18f0ecc80a69f8e0ae988991c39" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.317922 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvt4g" event={"ID":"74203cf3-9c19-4bd3-9d1b-040a99893ab1","Type":"ContainerDied","Data":"d131d9e6632780f6617d3c79c311eb78fa03ca587fd78fb79369e9fc8213045b"} Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.317992 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvt4g" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.343289 5018 scope.go:117] "RemoveContainer" containerID="52dcaa037cf33580f3d267e9a6c30d361bf74086c2ef26b99fdae66386a3e70f" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.349938 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" podStartSLOduration=1.095811085 podStartE2EDuration="4.34992517s" podCreationTimestamp="2025-10-14 07:02:30 +0000 UTC" firstStartedPulling="2025-10-14 07:02:30.903014221 +0000 UTC m=+767.487060848" lastFinishedPulling="2025-10-14 07:02:34.157128306 +0000 UTC m=+770.741174933" observedRunningTime="2025-10-14 07:02:34.332556145 +0000 UTC m=+770.916602772" watchObservedRunningTime="2025-10-14 07:02:34.34992517 +0000 UTC m=+770.933971797" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.366349 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.370544 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cvt4g"] Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.373837 5018 scope.go:117] "RemoveContainer" containerID="870aa281926e5f30fef860758d82c17a71cd7f846b8073c0bbb2803332f99cff" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.398886 5018 scope.go:117] "RemoveContainer" containerID="b1bdff870eaa3fca906238bccdeca4044ceab671e0aa358171497acb81201857" Oct 14 07:02:34 crc kubenswrapper[5018]: I1014 07:02:34.632158 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" path="/var/lib/kubelet/pods/74203cf3-9c19-4bd3-9d1b-040a99893ab1/volumes" Oct 14 07:02:35 crc kubenswrapper[5018]: I1014 07:02:35.332756 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-556sb" event={"ID":"c32afd97-261a-4e41-bda2-ae5de91f5d6f","Type":"ContainerStarted","Data":"6c26c8eb6383f82b732acf31bf71b5f10beed79d45f355fe49fb3427ab68cf42"} Oct 14 07:02:35 crc kubenswrapper[5018]: I1014 07:02:35.333280 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:35 crc kubenswrapper[5018]: I1014 07:02:35.345570 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" event={"ID":"b6eb676d-62fd-4035-8cfd-4617d1cb7b35","Type":"ContainerStarted","Data":"3d777fd4ffd22b7cd9a77c40535d6e388c246daa17432b6b3cd5790585d4c5b2"} Oct 14 07:02:35 crc kubenswrapper[5018]: I1014 07:02:35.367789 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-556sb" podStartSLOduration=1.890706006 podStartE2EDuration="5.367726647s" podCreationTimestamp="2025-10-14 07:02:30 +0000 UTC" firstStartedPulling="2025-10-14 07:02:30.64877246 +0000 UTC m=+767.232819087" lastFinishedPulling="2025-10-14 07:02:34.125793091 +0000 UTC m=+770.709839728" observedRunningTime="2025-10-14 07:02:35.365237968 +0000 UTC m=+771.949284625" watchObservedRunningTime="2025-10-14 07:02:35.367726647 +0000 UTC m=+771.951773304" Oct 14 07:02:35 crc kubenswrapper[5018]: I1014 07:02:35.390827 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-gbl7v" podStartSLOduration=2.4060381 podStartE2EDuration="5.390808692s" podCreationTimestamp="2025-10-14 07:02:30 +0000 UTC" firstStartedPulling="2025-10-14 07:02:31.13606478 +0000 UTC m=+767.720111457" lastFinishedPulling="2025-10-14 07:02:34.120835412 +0000 UTC m=+770.704882049" observedRunningTime="2025-10-14 07:02:35.387680104 +0000 UTC m=+771.971726721" watchObservedRunningTime="2025-10-14 07:02:35.390808692 +0000 UTC m=+771.974855329" Oct 14 07:02:37 crc kubenswrapper[5018]: I1014 07:02:37.365303 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" event={"ID":"c718fdd6-a84c-4842-aa5d-f6ccc0c8babd","Type":"ContainerStarted","Data":"616f94284f00a4c201552ffd0e171197277ca630e3e1afa85a8b930fa61cd646"} Oct 14 07:02:37 crc kubenswrapper[5018]: I1014 07:02:37.395186 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lszjp" podStartSLOduration=1.9379575949999999 podStartE2EDuration="7.395155731s" podCreationTimestamp="2025-10-14 07:02:30 +0000 UTC" firstStartedPulling="2025-10-14 07:02:31.000551515 +0000 UTC m=+767.584598152" lastFinishedPulling="2025-10-14 07:02:36.457749661 +0000 UTC m=+773.041796288" observedRunningTime="2025-10-14 07:02:37.38973648 +0000 UTC m=+773.973783147" watchObservedRunningTime="2025-10-14 07:02:37.395155731 +0000 UTC m=+773.979202388" Oct 14 07:02:40 crc kubenswrapper[5018]: I1014 07:02:40.657388 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-556sb" Oct 14 07:02:40 crc kubenswrapper[5018]: I1014 07:02:40.926240 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:40 crc kubenswrapper[5018]: I1014 07:02:40.926344 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:40 crc kubenswrapper[5018]: I1014 07:02:40.934200 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:41 crc kubenswrapper[5018]: I1014 07:02:41.397112 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-bddbd9cfb-t9fh7" Oct 14 07:02:41 crc kubenswrapper[5018]: I1014 07:02:41.460098 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 07:02:50 crc kubenswrapper[5018]: I1014 07:02:50.593840 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-rcf8x" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.467086 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:02:55 crc kubenswrapper[5018]: E1014 07:02:55.468265 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="registry-server" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.468294 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="registry-server" Oct 14 07:02:55 crc kubenswrapper[5018]: E1014 07:02:55.468317 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="extract-content" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.468329 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="extract-content" Oct 14 07:02:55 crc kubenswrapper[5018]: E1014 07:02:55.468371 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="extract-utilities" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.468386 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="extract-utilities" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.468561 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="74203cf3-9c19-4bd3-9d1b-040a99893ab1" containerName="registry-server" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.469984 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.481696 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.604260 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.604302 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.604336 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlb2d\" (UniqueName: \"kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.705130 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.705171 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.705203 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlb2d\" (UniqueName: \"kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.705689 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.705916 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.736264 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlb2d\" (UniqueName: \"kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d\") pod \"certified-operators-tqgn4\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:55 crc kubenswrapper[5018]: I1014 07:02:55.801559 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:02:56 crc kubenswrapper[5018]: I1014 07:02:56.296528 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:02:56 crc kubenswrapper[5018]: I1014 07:02:56.514457 5018 generic.go:334] "Generic (PLEG): container finished" podID="f754eaca-403f-4e40-b306-acc795471f22" containerID="5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca" exitCode=0 Oct 14 07:02:56 crc kubenswrapper[5018]: I1014 07:02:56.514536 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerDied","Data":"5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca"} Oct 14 07:02:56 crc kubenswrapper[5018]: I1014 07:02:56.514949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerStarted","Data":"e60ee11228e21316aae8f860a0f3666e1822f42ac8be4896cb82ab0f264d43bf"} Oct 14 07:02:58 crc kubenswrapper[5018]: I1014 07:02:58.534581 5018 generic.go:334] "Generic (PLEG): container finished" podID="f754eaca-403f-4e40-b306-acc795471f22" containerID="d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42" exitCode=0 Oct 14 07:02:58 crc kubenswrapper[5018]: I1014 07:02:58.535026 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerDied","Data":"d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42"} Oct 14 07:02:59 crc kubenswrapper[5018]: I1014 07:02:59.542674 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerStarted","Data":"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c"} Oct 14 07:02:59 crc kubenswrapper[5018]: I1014 07:02:59.564889 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tqgn4" podStartSLOduration=2.071904124 podStartE2EDuration="4.56485s" podCreationTimestamp="2025-10-14 07:02:55 +0000 UTC" firstStartedPulling="2025-10-14 07:02:56.515930635 +0000 UTC m=+793.099977272" lastFinishedPulling="2025-10-14 07:02:59.008876481 +0000 UTC m=+795.592923148" observedRunningTime="2025-10-14 07:02:59.556925276 +0000 UTC m=+796.140971923" watchObservedRunningTime="2025-10-14 07:02:59.56485 +0000 UTC m=+796.148896627" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.060149 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.062603 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.071090 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.216040 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rvln\" (UniqueName: \"kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.216108 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.216162 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.317685 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rvln\" (UniqueName: \"kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.318081 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.318165 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.318865 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.319079 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.351108 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rvln\" (UniqueName: \"kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln\") pod \"community-operators-7qs6r\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.392190 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:03 crc kubenswrapper[5018]: I1014 07:03:03.892772 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:03 crc kubenswrapper[5018]: W1014 07:03:03.902873 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod941ac9be_ce77_449e_957d_a82cf63071c5.slice/crio-08f2bd7da23e1d3edfdbb55810fd407ffb87fe4270c874786528434907c20029 WatchSource:0}: Error finding container 08f2bd7da23e1d3edfdbb55810fd407ffb87fe4270c874786528434907c20029: Status 404 returned error can't find the container with id 08f2bd7da23e1d3edfdbb55810fd407ffb87fe4270c874786528434907c20029 Oct 14 07:03:04 crc kubenswrapper[5018]: I1014 07:03:04.575975 5018 generic.go:334] "Generic (PLEG): container finished" podID="941ac9be-ce77-449e-957d-a82cf63071c5" containerID="42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3" exitCode=0 Oct 14 07:03:04 crc kubenswrapper[5018]: I1014 07:03:04.576328 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerDied","Data":"42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3"} Oct 14 07:03:04 crc kubenswrapper[5018]: I1014 07:03:04.576367 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerStarted","Data":"08f2bd7da23e1d3edfdbb55810fd407ffb87fe4270c874786528434907c20029"} Oct 14 07:03:05 crc kubenswrapper[5018]: I1014 07:03:05.803135 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:05 crc kubenswrapper[5018]: I1014 07:03:05.803548 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:05 crc kubenswrapper[5018]: I1014 07:03:05.851887 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:06 crc kubenswrapper[5018]: I1014 07:03:06.513877 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sc99m" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" containerID="cri-o://266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd" gracePeriod=15 Oct 14 07:03:06 crc kubenswrapper[5018]: I1014 07:03:06.637848 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:06 crc kubenswrapper[5018]: I1014 07:03:06.945297 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sc99m_83a99049-d48d-4e15-bb4d-3eb09e9eb971/console/0.log" Oct 14 07:03:06 crc kubenswrapper[5018]: I1014 07:03:06.945363 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073338 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073468 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073521 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073577 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gms6\" (UniqueName: \"kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073666 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073704 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073776 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca\") pod \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\" (UID: \"83a99049-d48d-4e15-bb4d-3eb09e9eb971\") " Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.073916 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config" (OuterVolumeSpecName: "console-config") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.074133 5018 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.074743 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.074794 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.075007 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca" (OuterVolumeSpecName: "service-ca") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.080749 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.081128 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.088468 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6" (OuterVolumeSpecName: "kube-api-access-2gms6") pod "83a99049-d48d-4e15-bb4d-3eb09e9eb971" (UID: "83a99049-d48d-4e15-bb4d-3eb09e9eb971"). InnerVolumeSpecName "kube-api-access-2gms6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.174898 5018 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.175223 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gms6\" (UniqueName: \"kubernetes.io/projected/83a99049-d48d-4e15-bb4d-3eb09e9eb971-kube-api-access-2gms6\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.175238 5018 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.175251 5018 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.175263 5018 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/83a99049-d48d-4e15-bb4d-3eb09e9eb971-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.175274 5018 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/83a99049-d48d-4e15-bb4d-3eb09e9eb971-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.427817 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.595536 5018 generic.go:334] "Generic (PLEG): container finished" podID="941ac9be-ce77-449e-957d-a82cf63071c5" containerID="159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210" exitCode=0 Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.595609 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerDied","Data":"159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210"} Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597390 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sc99m_83a99049-d48d-4e15-bb4d-3eb09e9eb971/console/0.log" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597423 5018 generic.go:334] "Generic (PLEG): container finished" podID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerID="266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd" exitCode=2 Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597511 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sc99m" event={"ID":"83a99049-d48d-4e15-bb4d-3eb09e9eb971","Type":"ContainerDied","Data":"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd"} Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597552 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sc99m" event={"ID":"83a99049-d48d-4e15-bb4d-3eb09e9eb971","Type":"ContainerDied","Data":"4cf80d3661af2a3d6001824a3ade4db10d002607f7169d20ec92f88323133f75"} Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597594 5018 scope.go:117] "RemoveContainer" containerID="266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.597860 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sc99m" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.623480 5018 scope.go:117] "RemoveContainer" containerID="266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd" Oct 14 07:03:07 crc kubenswrapper[5018]: E1014 07:03:07.624107 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd\": container with ID starting with 266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd not found: ID does not exist" containerID="266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.624159 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd"} err="failed to get container status \"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd\": rpc error: code = NotFound desc = could not find container \"266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd\": container with ID starting with 266da9391abef611ae5607099a8038ea4672b3da0da24737b50412112e49b3cd not found: ID does not exist" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.644721 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.648393 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sc99m"] Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.898865 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb"] Oct 14 07:03:07 crc kubenswrapper[5018]: E1014 07:03:07.899195 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.899214 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.899379 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" containerName="console" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.900556 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.903033 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:03:07 crc kubenswrapper[5018]: I1014 07:03:07.919332 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb"] Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.100772 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.100879 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5tbh\" (UniqueName: \"kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.101109 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.202517 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.202644 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5tbh\" (UniqueName: \"kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.202788 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.203388 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.203545 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.235478 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5tbh\" (UniqueName: \"kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.519528 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.621209 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a99049-d48d-4e15-bb4d-3eb09e9eb971" path="/var/lib/kubelet/pods/83a99049-d48d-4e15-bb4d-3eb09e9eb971/volumes" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.677565 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerStarted","Data":"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635"} Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.679102 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tqgn4" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="registry-server" containerID="cri-o://480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c" gracePeriod=2 Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.700448 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7qs6r" podStartSLOduration=2.035690852 podStartE2EDuration="5.700429562s" podCreationTimestamp="2025-10-14 07:03:03 +0000 UTC" firstStartedPulling="2025-10-14 07:03:04.578130738 +0000 UTC m=+801.162177395" lastFinishedPulling="2025-10-14 07:03:08.242869438 +0000 UTC m=+804.826916105" observedRunningTime="2025-10-14 07:03:08.699473755 +0000 UTC m=+805.283520402" watchObservedRunningTime="2025-10-14 07:03:08.700429562 +0000 UTC m=+805.284476189" Oct 14 07:03:08 crc kubenswrapper[5018]: I1014 07:03:08.996919 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb"] Oct 14 07:03:09 crc kubenswrapper[5018]: W1014 07:03:09.005674 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod282c68db_308a_4370_99fd_c20f6a5ff81d.slice/crio-c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f WatchSource:0}: Error finding container c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f: Status 404 returned error can't find the container with id c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.027159 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.121173 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities\") pod \"f754eaca-403f-4e40-b306-acc795471f22\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.121237 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content\") pod \"f754eaca-403f-4e40-b306-acc795471f22\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.121294 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlb2d\" (UniqueName: \"kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d\") pod \"f754eaca-403f-4e40-b306-acc795471f22\" (UID: \"f754eaca-403f-4e40-b306-acc795471f22\") " Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.122258 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities" (OuterVolumeSpecName: "utilities") pod "f754eaca-403f-4e40-b306-acc795471f22" (UID: "f754eaca-403f-4e40-b306-acc795471f22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.129794 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d" (OuterVolumeSpecName: "kube-api-access-hlb2d") pod "f754eaca-403f-4e40-b306-acc795471f22" (UID: "f754eaca-403f-4e40-b306-acc795471f22"). InnerVolumeSpecName "kube-api-access-hlb2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.163906 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f754eaca-403f-4e40-b306-acc795471f22" (UID: "f754eaca-403f-4e40-b306-acc795471f22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.222274 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.222304 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f754eaca-403f-4e40-b306-acc795471f22-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.222314 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlb2d\" (UniqueName: \"kubernetes.io/projected/f754eaca-403f-4e40-b306-acc795471f22-kube-api-access-hlb2d\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.684966 5018 generic.go:334] "Generic (PLEG): container finished" podID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerID="623096ae135bb63d6c5d866ef3dd85102261b6c74f2adc7703e18757c00310c3" exitCode=0 Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.685137 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerDied","Data":"623096ae135bb63d6c5d866ef3dd85102261b6c74f2adc7703e18757c00310c3"} Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.685470 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerStarted","Data":"c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f"} Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.690915 5018 generic.go:334] "Generic (PLEG): container finished" podID="f754eaca-403f-4e40-b306-acc795471f22" containerID="480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c" exitCode=0 Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.691486 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgn4" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.691791 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerDied","Data":"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c"} Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.691822 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgn4" event={"ID":"f754eaca-403f-4e40-b306-acc795471f22","Type":"ContainerDied","Data":"e60ee11228e21316aae8f860a0f3666e1822f42ac8be4896cb82ab0f264d43bf"} Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.691869 5018 scope.go:117] "RemoveContainer" containerID="480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.713247 5018 scope.go:117] "RemoveContainer" containerID="d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.726084 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.728814 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tqgn4"] Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.748874 5018 scope.go:117] "RemoveContainer" containerID="5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.766331 5018 scope.go:117] "RemoveContainer" containerID="480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c" Oct 14 07:03:09 crc kubenswrapper[5018]: E1014 07:03:09.766794 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c\": container with ID starting with 480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c not found: ID does not exist" containerID="480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.766824 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c"} err="failed to get container status \"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c\": rpc error: code = NotFound desc = could not find container \"480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c\": container with ID starting with 480fe4602d4510b4d4e5437c0b33a8b724b797003ef105fd8cf5c9f547ace81c not found: ID does not exist" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.766845 5018 scope.go:117] "RemoveContainer" containerID="d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42" Oct 14 07:03:09 crc kubenswrapper[5018]: E1014 07:03:09.767246 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42\": container with ID starting with d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42 not found: ID does not exist" containerID="d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.767438 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42"} err="failed to get container status \"d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42\": rpc error: code = NotFound desc = could not find container \"d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42\": container with ID starting with d8180bdf9f75ef2a0d477f796b180b412fefaf185d4f44273c175cf59f967f42 not found: ID does not exist" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.767592 5018 scope.go:117] "RemoveContainer" containerID="5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca" Oct 14 07:03:09 crc kubenswrapper[5018]: E1014 07:03:09.768019 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca\": container with ID starting with 5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca not found: ID does not exist" containerID="5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca" Oct 14 07:03:09 crc kubenswrapper[5018]: I1014 07:03:09.768043 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca"} err="failed to get container status \"5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca\": rpc error: code = NotFound desc = could not find container \"5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca\": container with ID starting with 5a4a994f8e0341b4462e291ee3e685de4a459a9e14ee7aa635b2a4d936a9deca not found: ID does not exist" Oct 14 07:03:10 crc kubenswrapper[5018]: I1014 07:03:10.616280 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f754eaca-403f-4e40-b306-acc795471f22" path="/var/lib/kubelet/pods/f754eaca-403f-4e40-b306-acc795471f22/volumes" Oct 14 07:03:11 crc kubenswrapper[5018]: I1014 07:03:11.706919 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerStarted","Data":"b319af4dedebcfc0cf9581d1aba412b50ba85d4b3346f1f923cf60f7ab17d8e5"} Oct 14 07:03:12 crc kubenswrapper[5018]: I1014 07:03:12.717166 5018 generic.go:334] "Generic (PLEG): container finished" podID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerID="b319af4dedebcfc0cf9581d1aba412b50ba85d4b3346f1f923cf60f7ab17d8e5" exitCode=0 Oct 14 07:03:12 crc kubenswrapper[5018]: I1014 07:03:12.717229 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerDied","Data":"b319af4dedebcfc0cf9581d1aba412b50ba85d4b3346f1f923cf60f7ab17d8e5"} Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.392874 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.393404 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.431585 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.729383 5018 generic.go:334] "Generic (PLEG): container finished" podID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerID="978043aa592e645dcb036847a3de5ff5efad704a1fcd85b166a9d82f6c007093" exitCode=0 Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.729448 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerDied","Data":"978043aa592e645dcb036847a3de5ff5efad704a1fcd85b166a9d82f6c007093"} Oct 14 07:03:13 crc kubenswrapper[5018]: I1014 07:03:13.785298 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.100311 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.205119 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util\") pod \"282c68db-308a-4370-99fd-c20f6a5ff81d\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.205185 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle\") pod \"282c68db-308a-4370-99fd-c20f6a5ff81d\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.205701 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5tbh\" (UniqueName: \"kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh\") pod \"282c68db-308a-4370-99fd-c20f6a5ff81d\" (UID: \"282c68db-308a-4370-99fd-c20f6a5ff81d\") " Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.209109 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle" (OuterVolumeSpecName: "bundle") pod "282c68db-308a-4370-99fd-c20f6a5ff81d" (UID: "282c68db-308a-4370-99fd-c20f6a5ff81d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.214606 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh" (OuterVolumeSpecName: "kube-api-access-s5tbh") pod "282c68db-308a-4370-99fd-c20f6a5ff81d" (UID: "282c68db-308a-4370-99fd-c20f6a5ff81d"). InnerVolumeSpecName "kube-api-access-s5tbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.215950 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util" (OuterVolumeSpecName: "util") pod "282c68db-308a-4370-99fd-c20f6a5ff81d" (UID: "282c68db-308a-4370-99fd-c20f6a5ff81d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.307438 5018 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.307476 5018 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/282c68db-308a-4370-99fd-c20f6a5ff81d-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.307490 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5tbh\" (UniqueName: \"kubernetes.io/projected/282c68db-308a-4370-99fd-c20f6a5ff81d-kube-api-access-s5tbh\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.749844 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" event={"ID":"282c68db-308a-4370-99fd-c20f6a5ff81d","Type":"ContainerDied","Data":"c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f"} Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.749896 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9abe790ec6f3a721ea096a918ce622511223a2e41d87fa8a4d1bbe455a3a67f" Oct 14 07:03:15 crc kubenswrapper[5018]: I1014 07:03:15.749973 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.046309 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.046886 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7qs6r" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="registry-server" containerID="cri-o://d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635" gracePeriod=2 Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.432345 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.634217 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content\") pod \"941ac9be-ce77-449e-957d-a82cf63071c5\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.634339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rvln\" (UniqueName: \"kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln\") pod \"941ac9be-ce77-449e-957d-a82cf63071c5\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.634406 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities\") pod \"941ac9be-ce77-449e-957d-a82cf63071c5\" (UID: \"941ac9be-ce77-449e-957d-a82cf63071c5\") " Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.635247 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities" (OuterVolumeSpecName: "utilities") pod "941ac9be-ce77-449e-957d-a82cf63071c5" (UID: "941ac9be-ce77-449e-957d-a82cf63071c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.640986 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln" (OuterVolumeSpecName: "kube-api-access-7rvln") pod "941ac9be-ce77-449e-957d-a82cf63071c5" (UID: "941ac9be-ce77-449e-957d-a82cf63071c5"). InnerVolumeSpecName "kube-api-access-7rvln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.735675 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rvln\" (UniqueName: \"kubernetes.io/projected/941ac9be-ce77-449e-957d-a82cf63071c5-kube-api-access-7rvln\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.735705 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.776819 5018 generic.go:334] "Generic (PLEG): container finished" podID="941ac9be-ce77-449e-957d-a82cf63071c5" containerID="d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635" exitCode=0 Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.776870 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerDied","Data":"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635"} Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.776914 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qs6r" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.776935 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qs6r" event={"ID":"941ac9be-ce77-449e-957d-a82cf63071c5","Type":"ContainerDied","Data":"08f2bd7da23e1d3edfdbb55810fd407ffb87fe4270c874786528434907c20029"} Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.776956 5018 scope.go:117] "RemoveContainer" containerID="d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.791568 5018 scope.go:117] "RemoveContainer" containerID="159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.807990 5018 scope.go:117] "RemoveContainer" containerID="42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.810942 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "941ac9be-ce77-449e-957d-a82cf63071c5" (UID: "941ac9be-ce77-449e-957d-a82cf63071c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.833840 5018 scope.go:117] "RemoveContainer" containerID="d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635" Oct 14 07:03:17 crc kubenswrapper[5018]: E1014 07:03:17.834341 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635\": container with ID starting with d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635 not found: ID does not exist" containerID="d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.834377 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635"} err="failed to get container status \"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635\": rpc error: code = NotFound desc = could not find container \"d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635\": container with ID starting with d5dd41931645be5a4ba7df556ca95143936f988f52350285f409e7a72de17635 not found: ID does not exist" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.834397 5018 scope.go:117] "RemoveContainer" containerID="159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210" Oct 14 07:03:17 crc kubenswrapper[5018]: E1014 07:03:17.834742 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210\": container with ID starting with 159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210 not found: ID does not exist" containerID="159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.834766 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210"} err="failed to get container status \"159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210\": rpc error: code = NotFound desc = could not find container \"159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210\": container with ID starting with 159f66f045fec65fe0925c5bb5745bc7f8c854b786dd936ee5435a552540d210 not found: ID does not exist" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.834782 5018 scope.go:117] "RemoveContainer" containerID="42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3" Oct 14 07:03:17 crc kubenswrapper[5018]: E1014 07:03:17.835099 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3\": container with ID starting with 42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3 not found: ID does not exist" containerID="42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.835169 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3"} err="failed to get container status \"42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3\": rpc error: code = NotFound desc = could not find container \"42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3\": container with ID starting with 42d71dfc4f4be01bc9707ba14ccea9989a9a873f0cc663acbfac2ed9345b0ea3 not found: ID does not exist" Oct 14 07:03:17 crc kubenswrapper[5018]: I1014 07:03:17.837014 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/941ac9be-ce77-449e-957d-a82cf63071c5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:18 crc kubenswrapper[5018]: I1014 07:03:18.121642 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:18 crc kubenswrapper[5018]: I1014 07:03:18.131641 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7qs6r"] Oct 14 07:03:18 crc kubenswrapper[5018]: I1014 07:03:18.612317 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" path="/var/lib/kubelet/pods/941ac9be-ce77-449e-957d-a82cf63071c5/volumes" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.232537 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233108 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="extract-content" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233122 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="extract-content" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233138 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="pull" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233145 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="pull" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233158 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233167 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233178 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="util" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233185 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="util" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233197 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="extract-utilities" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233204 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="extract-utilities" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233213 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="extract" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233220 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="extract" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233231 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="extract-utilities" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233238 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="extract-utilities" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233247 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233254 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: E1014 07:03:21.233263 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="extract-content" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233270 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="extract-content" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233388 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="941ac9be-ce77-449e-957d-a82cf63071c5" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233406 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f754eaca-403f-4e40-b306-acc795471f22" containerName="registry-server" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.233417 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="282c68db-308a-4370-99fd-c20f6a5ff81d" containerName="extract" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.234198 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.243565 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.375303 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.375354 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.375378 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zx6r\" (UniqueName: \"kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.477269 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.477384 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.477440 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zx6r\" (UniqueName: \"kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.477959 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.478089 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.507838 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zx6r\" (UniqueName: \"kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r\") pod \"redhat-marketplace-vx86j\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:21 crc kubenswrapper[5018]: I1014 07:03:21.588490 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:22 crc kubenswrapper[5018]: I1014 07:03:22.047098 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:22 crc kubenswrapper[5018]: I1014 07:03:22.804287 5018 generic.go:334] "Generic (PLEG): container finished" podID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerID="9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730" exitCode=0 Oct 14 07:03:22 crc kubenswrapper[5018]: I1014 07:03:22.804362 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerDied","Data":"9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730"} Oct 14 07:03:22 crc kubenswrapper[5018]: I1014 07:03:22.804647 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerStarted","Data":"0fa5877886b02c68caa41ab45a26890cd6f600f811c829b438ebe2b0422ca759"} Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.812512 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerStarted","Data":"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05"} Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.892187 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj"] Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.894916 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.905112 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.905734 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.909705 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.910397 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-webhook-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.910443 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-apiservice-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.910522 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lmbb\" (UniqueName: \"kubernetes.io/projected/5fb208e9-58f7-43be-81d4-ecfa72a34dba-kube-api-access-4lmbb\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.910871 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-t9fnn" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.914874 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 14 07:03:23 crc kubenswrapper[5018]: I1014 07:03:23.930699 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj"] Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.011267 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-webhook-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.011319 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-apiservice-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.011380 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lmbb\" (UniqueName: \"kubernetes.io/projected/5fb208e9-58f7-43be-81d4-ecfa72a34dba-kube-api-access-4lmbb\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.019532 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-apiservice-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.030284 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5fb208e9-58f7-43be-81d4-ecfa72a34dba-webhook-cert\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.034521 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lmbb\" (UniqueName: \"kubernetes.io/projected/5fb208e9-58f7-43be-81d4-ecfa72a34dba-kube-api-access-4lmbb\") pod \"metallb-operator-controller-manager-6f688cc4bd-dzcxj\" (UID: \"5fb208e9-58f7-43be-81d4-ecfa72a34dba\") " pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.124421 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6784749796-zrctk"] Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.125271 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.127069 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.127435 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jxpnq" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.127413 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.145936 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6784749796-zrctk"] Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.214335 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77dkj\" (UniqueName: \"kubernetes.io/projected/62f6d4e3-ca4f-44c8-9206-eda31a949201-kube-api-access-77dkj\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.214608 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-apiservice-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.214764 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-webhook-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.219129 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.316326 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-apiservice-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.316391 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-webhook-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.316430 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77dkj\" (UniqueName: \"kubernetes.io/projected/62f6d4e3-ca4f-44c8-9206-eda31a949201-kube-api-access-77dkj\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.323915 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-webhook-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.324384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62f6d4e3-ca4f-44c8-9206-eda31a949201-apiservice-cert\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.342473 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77dkj\" (UniqueName: \"kubernetes.io/projected/62f6d4e3-ca4f-44c8-9206-eda31a949201-kube-api-access-77dkj\") pod \"metallb-operator-webhook-server-6784749796-zrctk\" (UID: \"62f6d4e3-ca4f-44c8-9206-eda31a949201\") " pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.445452 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.727669 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj"] Oct 14 07:03:24 crc kubenswrapper[5018]: W1014 07:03:24.774957 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fb208e9_58f7_43be_81d4_ecfa72a34dba.slice/crio-d82e8fcbbbc34c236585391e36a4da26c1eaa3398672832e6c3b472018f01d53 WatchSource:0}: Error finding container d82e8fcbbbc34c236585391e36a4da26c1eaa3398672832e6c3b472018f01d53: Status 404 returned error can't find the container with id d82e8fcbbbc34c236585391e36a4da26c1eaa3398672832e6c3b472018f01d53 Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.804158 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6784749796-zrctk"] Oct 14 07:03:24 crc kubenswrapper[5018]: W1014 07:03:24.812356 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62f6d4e3_ca4f_44c8_9206_eda31a949201.slice/crio-89e19eca5dbf08e2061ca4840e77270eeb6ead9f790b4586d2553077d194c7ae WatchSource:0}: Error finding container 89e19eca5dbf08e2061ca4840e77270eeb6ead9f790b4586d2553077d194c7ae: Status 404 returned error can't find the container with id 89e19eca5dbf08e2061ca4840e77270eeb6ead9f790b4586d2553077d194c7ae Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.823016 5018 generic.go:334] "Generic (PLEG): container finished" podID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerID="2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05" exitCode=0 Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.823064 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerDied","Data":"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05"} Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.825494 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" event={"ID":"62f6d4e3-ca4f-44c8-9206-eda31a949201","Type":"ContainerStarted","Data":"89e19eca5dbf08e2061ca4840e77270eeb6ead9f790b4586d2553077d194c7ae"} Oct 14 07:03:24 crc kubenswrapper[5018]: I1014 07:03:24.826724 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" event={"ID":"5fb208e9-58f7-43be-81d4-ecfa72a34dba","Type":"ContainerStarted","Data":"d82e8fcbbbc34c236585391e36a4da26c1eaa3398672832e6c3b472018f01d53"} Oct 14 07:03:25 crc kubenswrapper[5018]: I1014 07:03:25.836538 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerStarted","Data":"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780"} Oct 14 07:03:25 crc kubenswrapper[5018]: I1014 07:03:25.860232 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vx86j" podStartSLOduration=2.377983477 podStartE2EDuration="4.860209809s" podCreationTimestamp="2025-10-14 07:03:21 +0000 UTC" firstStartedPulling="2025-10-14 07:03:22.806038375 +0000 UTC m=+819.390084992" lastFinishedPulling="2025-10-14 07:03:25.288264697 +0000 UTC m=+821.872311324" observedRunningTime="2025-10-14 07:03:25.857548344 +0000 UTC m=+822.441594981" watchObservedRunningTime="2025-10-14 07:03:25.860209809 +0000 UTC m=+822.444256456" Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.869544 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" event={"ID":"5fb208e9-58f7-43be-81d4-ecfa72a34dba","Type":"ContainerStarted","Data":"374bb9e63298bb8995e3118fd00fca1ac49ea9d1e91eaf7ed9165bf137e61142"} Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.870465 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.873086 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" event={"ID":"62f6d4e3-ca4f-44c8-9206-eda31a949201","Type":"ContainerStarted","Data":"ed1df02ab01824ac58e51fd4e25ab62a2bd3feb70323bb82a66e8e8007b10edb"} Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.873239 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.901827 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" podStartSLOduration=2.899777598 podStartE2EDuration="7.901803777s" podCreationTimestamp="2025-10-14 07:03:23 +0000 UTC" firstStartedPulling="2025-10-14 07:03:24.77831083 +0000 UTC m=+821.362357467" lastFinishedPulling="2025-10-14 07:03:29.780337009 +0000 UTC m=+826.364383646" observedRunningTime="2025-10-14 07:03:30.901148449 +0000 UTC m=+827.485195116" watchObservedRunningTime="2025-10-14 07:03:30.901803777 +0000 UTC m=+827.485850434" Oct 14 07:03:30 crc kubenswrapper[5018]: I1014 07:03:30.939540 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" podStartSLOduration=1.962640538 podStartE2EDuration="6.939519545s" podCreationTimestamp="2025-10-14 07:03:24 +0000 UTC" firstStartedPulling="2025-10-14 07:03:24.819147106 +0000 UTC m=+821.403193733" lastFinishedPulling="2025-10-14 07:03:29.796026073 +0000 UTC m=+826.380072740" observedRunningTime="2025-10-14 07:03:30.935954154 +0000 UTC m=+827.520000821" watchObservedRunningTime="2025-10-14 07:03:30.939519545 +0000 UTC m=+827.523566202" Oct 14 07:03:31 crc kubenswrapper[5018]: I1014 07:03:31.589526 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:31 crc kubenswrapper[5018]: I1014 07:03:31.589605 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:31 crc kubenswrapper[5018]: I1014 07:03:31.639194 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:31 crc kubenswrapper[5018]: I1014 07:03:31.926209 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:32 crc kubenswrapper[5018]: I1014 07:03:32.430024 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:33 crc kubenswrapper[5018]: I1014 07:03:33.888526 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vx86j" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="registry-server" containerID="cri-o://4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780" gracePeriod=2 Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.282517 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.451934 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content\") pod \"98007c19-5d1f-4119-a3e6-c8768af848fb\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.452016 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zx6r\" (UniqueName: \"kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r\") pod \"98007c19-5d1f-4119-a3e6-c8768af848fb\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.452082 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities\") pod \"98007c19-5d1f-4119-a3e6-c8768af848fb\" (UID: \"98007c19-5d1f-4119-a3e6-c8768af848fb\") " Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.453445 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities" (OuterVolumeSpecName: "utilities") pod "98007c19-5d1f-4119-a3e6-c8768af848fb" (UID: "98007c19-5d1f-4119-a3e6-c8768af848fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.460847 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r" (OuterVolumeSpecName: "kube-api-access-6zx6r") pod "98007c19-5d1f-4119-a3e6-c8768af848fb" (UID: "98007c19-5d1f-4119-a3e6-c8768af848fb"). InnerVolumeSpecName "kube-api-access-6zx6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.464701 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98007c19-5d1f-4119-a3e6-c8768af848fb" (UID: "98007c19-5d1f-4119-a3e6-c8768af848fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.553745 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.553786 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98007c19-5d1f-4119-a3e6-c8768af848fb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.553803 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zx6r\" (UniqueName: \"kubernetes.io/projected/98007c19-5d1f-4119-a3e6-c8768af848fb-kube-api-access-6zx6r\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.899542 5018 generic.go:334] "Generic (PLEG): container finished" podID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerID="4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780" exitCode=0 Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.899595 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerDied","Data":"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780"} Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.899643 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx86j" event={"ID":"98007c19-5d1f-4119-a3e6-c8768af848fb","Type":"ContainerDied","Data":"0fa5877886b02c68caa41ab45a26890cd6f600f811c829b438ebe2b0422ca759"} Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.899656 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx86j" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.899663 5018 scope.go:117] "RemoveContainer" containerID="4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.920052 5018 scope.go:117] "RemoveContainer" containerID="2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.924432 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.937461 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx86j"] Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.938061 5018 scope.go:117] "RemoveContainer" containerID="9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.953458 5018 scope.go:117] "RemoveContainer" containerID="4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780" Oct 14 07:03:34 crc kubenswrapper[5018]: E1014 07:03:34.954207 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780\": container with ID starting with 4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780 not found: ID does not exist" containerID="4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.954265 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780"} err="failed to get container status \"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780\": rpc error: code = NotFound desc = could not find container \"4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780\": container with ID starting with 4207a19005c78204038408edb8a5d49a5a714c8f627d07f0e8889dd248b64780 not found: ID does not exist" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.954298 5018 scope.go:117] "RemoveContainer" containerID="2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05" Oct 14 07:03:34 crc kubenswrapper[5018]: E1014 07:03:34.954655 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05\": container with ID starting with 2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05 not found: ID does not exist" containerID="2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.954712 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05"} err="failed to get container status \"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05\": rpc error: code = NotFound desc = could not find container \"2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05\": container with ID starting with 2644b07ae302c1ec932088fc4b79f115ab7992fcf336918ea51ea311dd507b05 not found: ID does not exist" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.954750 5018 scope.go:117] "RemoveContainer" containerID="9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730" Oct 14 07:03:34 crc kubenswrapper[5018]: E1014 07:03:34.954981 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730\": container with ID starting with 9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730 not found: ID does not exist" containerID="9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730" Oct 14 07:03:34 crc kubenswrapper[5018]: I1014 07:03:34.955008 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730"} err="failed to get container status \"9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730\": rpc error: code = NotFound desc = could not find container \"9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730\": container with ID starting with 9dae3b4e4891fd2d6378963e0456fe1bd9b24e8909c63fb3337c042ef9093730 not found: ID does not exist" Oct 14 07:03:36 crc kubenswrapper[5018]: I1014 07:03:36.614873 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" path="/var/lib/kubelet/pods/98007c19-5d1f-4119-a3e6-c8768af848fb/volumes" Oct 14 07:03:44 crc kubenswrapper[5018]: I1014 07:03:44.454091 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6784749796-zrctk" Oct 14 07:04:04 crc kubenswrapper[5018]: I1014 07:04:04.222279 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6f688cc4bd-dzcxj" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.117056 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf"] Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.117609 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="registry-server" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.117643 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="registry-server" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.117660 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="extract-utilities" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.117668 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="extract-utilities" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.117679 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="extract-content" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.117688 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="extract-content" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.117809 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="98007c19-5d1f-4119-a3e6-c8768af848fb" containerName="registry-server" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.118246 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.120279 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7wh4b" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.120375 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.134533 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sv8bv"] Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.136661 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.138488 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.138673 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.167651 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf"] Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.196315 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9k9tt"] Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.197091 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.198878 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.199096 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dgjmr" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.199235 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.199344 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.225871 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-2xckb"] Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.226668 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.228827 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.253435 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.253497 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lc8s\" (UniqueName: \"kubernetes.io/projected/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-kube-api-access-5lc8s\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.253572 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-sockets\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.253983 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.254028 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-conf\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.254282 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-reloader\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.254368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-startup\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.254413 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rt4d\" (UniqueName: \"kubernetes.io/projected/db040ce3-0234-42ce-a316-9c05abc7dda5-kube-api-access-9rt4d\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.254446 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.269850 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2xckb"] Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355701 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/05f5473d-ed86-4967-9785-062bc507742f-metallb-excludel2\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355761 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-metrics-certs\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355789 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-startup\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355820 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rt4d\" (UniqueName: \"kubernetes.io/projected/db040ce3-0234-42ce-a316-9c05abc7dda5-kube-api-access-9rt4d\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.355875 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.356001 5018 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.356080 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert podName:d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d nodeName:}" failed. No retries permitted until 2025-10-14 07:04:05.856060149 +0000 UTC m=+862.440106776 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert") pod "frr-k8s-webhook-server-64bf5d555-84kzf" (UID: "d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d") : secret "frr-k8s-webhook-server-cert" not found Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356086 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqp9\" (UniqueName: \"kubernetes.io/projected/05f5473d-ed86-4967-9785-062bc507742f-kube-api-access-rkqp9\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356127 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-cert\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356156 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lc8s\" (UniqueName: \"kubernetes.io/projected/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-kube-api-access-5lc8s\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356181 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356215 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-sockets\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356247 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-metrics-certs\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356269 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356312 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-conf\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356349 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-reloader\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356359 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.356370 5018 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356373 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h87r\" (UniqueName: \"kubernetes.io/projected/b7499690-3aab-4d77-aa51-9381ab15492a-kube-api-access-8h87r\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.356440 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs podName:db040ce3-0234-42ce-a316-9c05abc7dda5 nodeName:}" failed. No retries permitted until 2025-10-14 07:04:05.856419829 +0000 UTC m=+862.440466456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs") pod "frr-k8s-sv8bv" (UID: "db040ce3-0234-42ce-a316-9c05abc7dda5") : secret "frr-k8s-certs-secret" not found Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356500 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-sockets\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356740 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-conf\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.356774 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/db040ce3-0234-42ce-a316-9c05abc7dda5-reloader\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.357316 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/db040ce3-0234-42ce-a316-9c05abc7dda5-frr-startup\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.376946 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lc8s\" (UniqueName: \"kubernetes.io/projected/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-kube-api-access-5lc8s\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.382901 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rt4d\" (UniqueName: \"kubernetes.io/projected/db040ce3-0234-42ce-a316-9c05abc7dda5-kube-api-access-9rt4d\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.456909 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-metrics-certs\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457029 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h87r\" (UniqueName: \"kubernetes.io/projected/b7499690-3aab-4d77-aa51-9381ab15492a-kube-api-access-8h87r\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457064 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/05f5473d-ed86-4967-9785-062bc507742f-metallb-excludel2\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457094 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-metrics-certs\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457158 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqp9\" (UniqueName: \"kubernetes.io/projected/05f5473d-ed86-4967-9785-062bc507742f-kube-api-access-rkqp9\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457205 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-cert\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.457236 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.457384 5018 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.457455 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist podName:05f5473d-ed86-4967-9785-062bc507742f nodeName:}" failed. No retries permitted until 2025-10-14 07:04:05.957432999 +0000 UTC m=+862.541479646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist") pod "speaker-9k9tt" (UID: "05f5473d-ed86-4967-9785-062bc507742f") : secret "metallb-memberlist" not found Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.458124 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/05f5473d-ed86-4967-9785-062bc507742f-metallb-excludel2\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.459481 5018 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.460828 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-metrics-certs\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.464095 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-metrics-certs\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.471940 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7499690-3aab-4d77-aa51-9381ab15492a-cert\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.476218 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqp9\" (UniqueName: \"kubernetes.io/projected/05f5473d-ed86-4967-9785-062bc507742f-kube-api-access-rkqp9\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.483092 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h87r\" (UniqueName: \"kubernetes.io/projected/b7499690-3aab-4d77-aa51-9381ab15492a-kube-api-access-8h87r\") pod \"controller-68d546b9d8-2xckb\" (UID: \"b7499690-3aab-4d77-aa51-9381ab15492a\") " pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.539219 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.862416 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.862964 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.867596 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-84kzf\" (UID: \"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.870485 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db040ce3-0234-42ce-a316-9c05abc7dda5-metrics-certs\") pod \"frr-k8s-sv8bv\" (UID: \"db040ce3-0234-42ce-a316-9c05abc7dda5\") " pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:05 crc kubenswrapper[5018]: I1014 07:04:05.963947 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.964136 5018 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 07:04:05 crc kubenswrapper[5018]: E1014 07:04:05.964217 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist podName:05f5473d-ed86-4967-9785-062bc507742f nodeName:}" failed. No retries permitted until 2025-10-14 07:04:06.964198026 +0000 UTC m=+863.548244663 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist") pod "speaker-9k9tt" (UID: "05f5473d-ed86-4967-9785-062bc507742f") : secret "metallb-memberlist" not found Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.030165 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.043170 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2xckb"] Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.050399 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:06 crc kubenswrapper[5018]: W1014 07:04:06.052853 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7499690_3aab_4d77_aa51_9381ab15492a.slice/crio-e01b0914bb41c3cf319a3ca260a7afd296a66bd55073bdbb74f2af5d188c5f75 WatchSource:0}: Error finding container e01b0914bb41c3cf319a3ca260a7afd296a66bd55073bdbb74f2af5d188c5f75: Status 404 returned error can't find the container with id e01b0914bb41c3cf319a3ca260a7afd296a66bd55073bdbb74f2af5d188c5f75 Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.103667 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2xckb" event={"ID":"b7499690-3aab-4d77-aa51-9381ab15492a","Type":"ContainerStarted","Data":"e01b0914bb41c3cf319a3ca260a7afd296a66bd55073bdbb74f2af5d188c5f75"} Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.512474 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf"] Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.976598 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:06 crc kubenswrapper[5018]: I1014 07:04:06.987530 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/05f5473d-ed86-4967-9785-062bc507742f-memberlist\") pod \"speaker-9k9tt\" (UID: \"05f5473d-ed86-4967-9785-062bc507742f\") " pod="metallb-system/speaker-9k9tt" Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.010403 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9k9tt" Oct 14 07:04:07 crc kubenswrapper[5018]: W1014 07:04:07.038774 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05f5473d_ed86_4967_9785_062bc507742f.slice/crio-0c6c4e7f9ddc68cab2b62c0c25e296d13f6fda6a282ea0af26f1738ec5fc9d09 WatchSource:0}: Error finding container 0c6c4e7f9ddc68cab2b62c0c25e296d13f6fda6a282ea0af26f1738ec5fc9d09: Status 404 returned error can't find the container with id 0c6c4e7f9ddc68cab2b62c0c25e296d13f6fda6a282ea0af26f1738ec5fc9d09 Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.113306 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k9tt" event={"ID":"05f5473d-ed86-4967-9785-062bc507742f","Type":"ContainerStarted","Data":"0c6c4e7f9ddc68cab2b62c0c25e296d13f6fda6a282ea0af26f1738ec5fc9d09"} Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.114983 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" event={"ID":"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d","Type":"ContainerStarted","Data":"5c63301adec99493c414492ce827693d1478fe3b93cdfb1fe64ed7191cf95394"} Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.116849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"f5d54364c339dede059fc5da9f7eb50d5de01362d1a295518900e110409f960c"} Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.120847 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2xckb" event={"ID":"b7499690-3aab-4d77-aa51-9381ab15492a","Type":"ContainerStarted","Data":"e345a2954f6c9200b320bba9c21eac527c2215968b227a602f1081565b1ca17e"} Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.120877 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2xckb" event={"ID":"b7499690-3aab-4d77-aa51-9381ab15492a","Type":"ContainerStarted","Data":"f5b674ca857f39ef80f009fdba227be5b5afc468bcdf7d74cb51ac7a627c1a2f"} Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.121568 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:07 crc kubenswrapper[5018]: I1014 07:04:07.141417 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-2xckb" podStartSLOduration=2.141364501 podStartE2EDuration="2.141364501s" podCreationTimestamp="2025-10-14 07:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:04:07.139240931 +0000 UTC m=+863.723287598" watchObservedRunningTime="2025-10-14 07:04:07.141364501 +0000 UTC m=+863.725411138" Oct 14 07:04:08 crc kubenswrapper[5018]: I1014 07:04:08.126842 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k9tt" event={"ID":"05f5473d-ed86-4967-9785-062bc507742f","Type":"ContainerStarted","Data":"3a5461407464ab4ee5f0893e992ba28400842ed2c9ae335a80df034b866c011d"} Oct 14 07:04:08 crc kubenswrapper[5018]: I1014 07:04:08.127167 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k9tt" event={"ID":"05f5473d-ed86-4967-9785-062bc507742f","Type":"ContainerStarted","Data":"bc7ee0f02a37ad7c6720c4cda0d091c227aaf595d95124e111c9314d86a9aec1"} Oct 14 07:04:08 crc kubenswrapper[5018]: I1014 07:04:08.151673 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9k9tt" podStartSLOduration=3.151655723 podStartE2EDuration="3.151655723s" podCreationTimestamp="2025-10-14 07:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:04:08.14518664 +0000 UTC m=+864.729233267" watchObservedRunningTime="2025-10-14 07:04:08.151655723 +0000 UTC m=+864.735702350" Oct 14 07:04:09 crc kubenswrapper[5018]: I1014 07:04:09.132425 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9k9tt" Oct 14 07:04:14 crc kubenswrapper[5018]: I1014 07:04:14.165989 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" event={"ID":"d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d","Type":"ContainerStarted","Data":"d23222b95e5e5c697f42be4e0bb8d2850d5f1c20538bae8565042e22fcf493dc"} Oct 14 07:04:14 crc kubenswrapper[5018]: I1014 07:04:14.166469 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:14 crc kubenswrapper[5018]: I1014 07:04:14.169575 5018 generic.go:334] "Generic (PLEG): container finished" podID="db040ce3-0234-42ce-a316-9c05abc7dda5" containerID="94fcfa576b239da91a83ef0557bb2438273b6d359328d96e8095f1b64c7bb83b" exitCode=0 Oct 14 07:04:14 crc kubenswrapper[5018]: I1014 07:04:14.169655 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerDied","Data":"94fcfa576b239da91a83ef0557bb2438273b6d359328d96e8095f1b64c7bb83b"} Oct 14 07:04:14 crc kubenswrapper[5018]: I1014 07:04:14.193961 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" podStartSLOduration=2.260141154 podStartE2EDuration="9.193938851s" podCreationTimestamp="2025-10-14 07:04:05 +0000 UTC" firstStartedPulling="2025-10-14 07:04:06.515654007 +0000 UTC m=+863.099700634" lastFinishedPulling="2025-10-14 07:04:13.449451704 +0000 UTC m=+870.033498331" observedRunningTime="2025-10-14 07:04:14.189319361 +0000 UTC m=+870.773365998" watchObservedRunningTime="2025-10-14 07:04:14.193938851 +0000 UTC m=+870.777985488" Oct 14 07:04:15 crc kubenswrapper[5018]: I1014 07:04:15.176482 5018 generic.go:334] "Generic (PLEG): container finished" podID="db040ce3-0234-42ce-a316-9c05abc7dda5" containerID="c7929f3edbcaeaf532641765cae157c1c3d0b0ea97dac46ad1aa95fc04bd650d" exitCode=0 Oct 14 07:04:15 crc kubenswrapper[5018]: I1014 07:04:15.176578 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerDied","Data":"c7929f3edbcaeaf532641765cae157c1c3d0b0ea97dac46ad1aa95fc04bd650d"} Oct 14 07:04:16 crc kubenswrapper[5018]: I1014 07:04:16.186525 5018 generic.go:334] "Generic (PLEG): container finished" podID="db040ce3-0234-42ce-a316-9c05abc7dda5" containerID="609c9812dc35c89ff4a6db570332797e32c9f1f81ffdee30a5e52ee0050cdb9b" exitCode=0 Oct 14 07:04:16 crc kubenswrapper[5018]: I1014 07:04:16.186683 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerDied","Data":"609c9812dc35c89ff4a6db570332797e32c9f1f81ffdee30a5e52ee0050cdb9b"} Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.018744 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9k9tt" Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.198023 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"826415cbd8cf033aa8c53bb47e1bceda5b024224eb50004a23900acdb5581a7c"} Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.198065 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"6e3c7286ac27027ccaf787685e898a2f58b1cba732f76104430c7d2447c595ce"} Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.198078 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"71f6afe5b50ac6ee0324a8bb29f9c0f4ed718ce175b383903106c2595e0a3bdb"} Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.198089 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"2fa0692fe26fe5c61505fc894eb409e51a0ad6ded1a952a194e6de38829d5ac0"} Oct 14 07:04:17 crc kubenswrapper[5018]: I1014 07:04:17.198099 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"3c51822a6b6d8a3ee9530a4afd764bdbfe25fea5bc9fd663305315398c6302b8"} Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.210013 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sv8bv" event={"ID":"db040ce3-0234-42ce-a316-9c05abc7dda5","Type":"ContainerStarted","Data":"fa1d842fd1c4132c3f3c3ffdf652b8ef2859da89fbd0904cf531153aa4455fe4"} Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.210360 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.243764 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sv8bv" podStartSLOduration=6.014608064 podStartE2EDuration="13.243741963s" podCreationTimestamp="2025-10-14 07:04:05 +0000 UTC" firstStartedPulling="2025-10-14 07:04:06.213471812 +0000 UTC m=+862.797518439" lastFinishedPulling="2025-10-14 07:04:13.442605711 +0000 UTC m=+870.026652338" observedRunningTime="2025-10-14 07:04:18.238269978 +0000 UTC m=+874.822316645" watchObservedRunningTime="2025-10-14 07:04:18.243741963 +0000 UTC m=+874.827788610" Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.883884 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw"] Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.885537 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.890566 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:04:18 crc kubenswrapper[5018]: I1014 07:04:18.905492 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw"] Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.039131 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.039579 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmr8j\" (UniqueName: \"kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.039944 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.141703 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmr8j\" (UniqueName: \"kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.141759 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.141811 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.142310 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.142353 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.163857 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmr8j\" (UniqueName: \"kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.213489 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:19 crc kubenswrapper[5018]: I1014 07:04:19.621942 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw"] Oct 14 07:04:20 crc kubenswrapper[5018]: I1014 07:04:20.225349 5018 generic.go:334] "Generic (PLEG): container finished" podID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerID="45feec9438d06fce161efa302f285e5ec05453aa797c708d8aa9536a514cbf79" exitCode=0 Oct 14 07:04:20 crc kubenswrapper[5018]: I1014 07:04:20.225661 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" event={"ID":"363b573a-a6f7-4633-8eb5-4d0f804f546d","Type":"ContainerDied","Data":"45feec9438d06fce161efa302f285e5ec05453aa797c708d8aa9536a514cbf79"} Oct 14 07:04:20 crc kubenswrapper[5018]: I1014 07:04:20.225707 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" event={"ID":"363b573a-a6f7-4633-8eb5-4d0f804f546d","Type":"ContainerStarted","Data":"eb72192c21be3cc330691128447cfe81b8d91da5132240ddba8c6a1d80bf77a7"} Oct 14 07:04:21 crc kubenswrapper[5018]: I1014 07:04:21.051487 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:21 crc kubenswrapper[5018]: I1014 07:04:21.096771 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:23 crc kubenswrapper[5018]: I1014 07:04:23.259057 5018 generic.go:334] "Generic (PLEG): container finished" podID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerID="5ee965f64fae47ab353d9b42aaef27bbcb3f6791c0eac5a115844885a8c544cf" exitCode=0 Oct 14 07:04:23 crc kubenswrapper[5018]: I1014 07:04:23.259140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" event={"ID":"363b573a-a6f7-4633-8eb5-4d0f804f546d","Type":"ContainerDied","Data":"5ee965f64fae47ab353d9b42aaef27bbcb3f6791c0eac5a115844885a8c544cf"} Oct 14 07:04:24 crc kubenswrapper[5018]: I1014 07:04:24.271686 5018 generic.go:334] "Generic (PLEG): container finished" podID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerID="22c12ed6fcda776ca444b119d5981b9bec838ebdc80a6e56c0decbb0b7f35275" exitCode=0 Oct 14 07:04:24 crc kubenswrapper[5018]: I1014 07:04:24.271778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" event={"ID":"363b573a-a6f7-4633-8eb5-4d0f804f546d","Type":"ContainerDied","Data":"22c12ed6fcda776ca444b119d5981b9bec838ebdc80a6e56c0decbb0b7f35275"} Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.535369 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.544201 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-2xckb" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.733225 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmr8j\" (UniqueName: \"kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j\") pod \"363b573a-a6f7-4633-8eb5-4d0f804f546d\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.733320 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle\") pod \"363b573a-a6f7-4633-8eb5-4d0f804f546d\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.733378 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util\") pod \"363b573a-a6f7-4633-8eb5-4d0f804f546d\" (UID: \"363b573a-a6f7-4633-8eb5-4d0f804f546d\") " Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.736259 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle" (OuterVolumeSpecName: "bundle") pod "363b573a-a6f7-4633-8eb5-4d0f804f546d" (UID: "363b573a-a6f7-4633-8eb5-4d0f804f546d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.745213 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j" (OuterVolumeSpecName: "kube-api-access-xmr8j") pod "363b573a-a6f7-4633-8eb5-4d0f804f546d" (UID: "363b573a-a6f7-4633-8eb5-4d0f804f546d"). InnerVolumeSpecName "kube-api-access-xmr8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.761293 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util" (OuterVolumeSpecName: "util") pod "363b573a-a6f7-4633-8eb5-4d0f804f546d" (UID: "363b573a-a6f7-4633-8eb5-4d0f804f546d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.835654 5018 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.836175 5018 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/363b573a-a6f7-4633-8eb5-4d0f804f546d-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:25 crc kubenswrapper[5018]: I1014 07:04:25.836207 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmr8j\" (UniqueName: \"kubernetes.io/projected/363b573a-a6f7-4633-8eb5-4d0f804f546d-kube-api-access-xmr8j\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:26 crc kubenswrapper[5018]: I1014 07:04:26.037767 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-84kzf" Oct 14 07:04:26 crc kubenswrapper[5018]: I1014 07:04:26.057050 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sv8bv" Oct 14 07:04:26 crc kubenswrapper[5018]: I1014 07:04:26.290987 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" event={"ID":"363b573a-a6f7-4633-8eb5-4d0f804f546d","Type":"ContainerDied","Data":"eb72192c21be3cc330691128447cfe81b8d91da5132240ddba8c6a1d80bf77a7"} Oct 14 07:04:26 crc kubenswrapper[5018]: I1014 07:04:26.291024 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb72192c21be3cc330691128447cfe81b8d91da5132240ddba8c6a1d80bf77a7" Oct 14 07:04:26 crc kubenswrapper[5018]: I1014 07:04:26.291098 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.620309 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg"] Oct 14 07:04:31 crc kubenswrapper[5018]: E1014 07:04:31.620840 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="util" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.620859 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="util" Oct 14 07:04:31 crc kubenswrapper[5018]: E1014 07:04:31.620879 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="pull" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.620889 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="pull" Oct 14 07:04:31 crc kubenswrapper[5018]: E1014 07:04:31.620905 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="extract" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.620913 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="extract" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.621061 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="363b573a-a6f7-4633-8eb5-4d0f804f546d" containerName="extract" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.621504 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.627764 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.628372 5018 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-rsv2d" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.629796 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.645703 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg"] Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.711765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g828q\" (UniqueName: \"kubernetes.io/projected/b6e8dfa4-bc9e-4aa3-8547-411966f42c68-kube-api-access-g828q\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4bkbg\" (UID: \"b6e8dfa4-bc9e-4aa3-8547-411966f42c68\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.812957 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g828q\" (UniqueName: \"kubernetes.io/projected/b6e8dfa4-bc9e-4aa3-8547-411966f42c68-kube-api-access-g828q\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4bkbg\" (UID: \"b6e8dfa4-bc9e-4aa3-8547-411966f42c68\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.843949 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g828q\" (UniqueName: \"kubernetes.io/projected/b6e8dfa4-bc9e-4aa3-8547-411966f42c68-kube-api-access-g828q\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4bkbg\" (UID: \"b6e8dfa4-bc9e-4aa3-8547-411966f42c68\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" Oct 14 07:04:31 crc kubenswrapper[5018]: I1014 07:04:31.941175 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" Oct 14 07:04:32 crc kubenswrapper[5018]: I1014 07:04:32.392909 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg"] Oct 14 07:04:32 crc kubenswrapper[5018]: W1014 07:04:32.397723 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6e8dfa4_bc9e_4aa3_8547_411966f42c68.slice/crio-fb152b8f0ca822ae79f92075e7179e5058fdc3f52a889fef67f5df3535fbed92 WatchSource:0}: Error finding container fb152b8f0ca822ae79f92075e7179e5058fdc3f52a889fef67f5df3535fbed92: Status 404 returned error can't find the container with id fb152b8f0ca822ae79f92075e7179e5058fdc3f52a889fef67f5df3535fbed92 Oct 14 07:04:33 crc kubenswrapper[5018]: I1014 07:04:33.343889 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" event={"ID":"b6e8dfa4-bc9e-4aa3-8547-411966f42c68","Type":"ContainerStarted","Data":"fb152b8f0ca822ae79f92075e7179e5058fdc3f52a889fef67f5df3535fbed92"} Oct 14 07:04:39 crc kubenswrapper[5018]: I1014 07:04:39.380086 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" event={"ID":"b6e8dfa4-bc9e-4aa3-8547-411966f42c68","Type":"ContainerStarted","Data":"f93a29cfc275792800027ee13cfc149ba6a373b9011c0fcb49ee232483325833"} Oct 14 07:04:39 crc kubenswrapper[5018]: I1014 07:04:39.422444 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4bkbg" podStartSLOduration=1.6324862489999998 podStartE2EDuration="8.422423175s" podCreationTimestamp="2025-10-14 07:04:31 +0000 UTC" firstStartedPulling="2025-10-14 07:04:32.400180203 +0000 UTC m=+888.984226830" lastFinishedPulling="2025-10-14 07:04:39.190117129 +0000 UTC m=+895.774163756" observedRunningTime="2025-10-14 07:04:39.399033443 +0000 UTC m=+895.983080120" watchObservedRunningTime="2025-10-14 07:04:39.422423175 +0000 UTC m=+896.006469822" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.160378 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-sb596"] Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.161406 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.163782 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.164265 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.164667 5018 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-kkt5c" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.176944 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-sb596"] Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.256066 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jxrv\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-kube-api-access-2jxrv\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.256144 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-bound-sa-token\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.358049 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jxrv\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-kube-api-access-2jxrv\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.358466 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-bound-sa-token\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.381732 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-bound-sa-token\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.382178 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jxrv\" (UniqueName: \"kubernetes.io/projected/e7159c40-ec11-43d3-84b4-1b3696caf977-kube-api-access-2jxrv\") pod \"cert-manager-webhook-d969966f-sb596\" (UID: \"e7159c40-ec11-43d3-84b4-1b3696caf977\") " pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.478114 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:43 crc kubenswrapper[5018]: I1014 07:04:43.925983 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-sb596"] Oct 14 07:04:43 crc kubenswrapper[5018]: W1014 07:04:43.936510 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7159c40_ec11_43d3_84b4_1b3696caf977.slice/crio-2f6b327481ce1c5b46128acf3fd6bd719d49f8e1449315420c62efb062c67050 WatchSource:0}: Error finding container 2f6b327481ce1c5b46128acf3fd6bd719d49f8e1449315420c62efb062c67050: Status 404 returned error can't find the container with id 2f6b327481ce1c5b46128acf3fd6bd719d49f8e1449315420c62efb062c67050 Oct 14 07:04:44 crc kubenswrapper[5018]: I1014 07:04:44.407739 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-sb596" event={"ID":"e7159c40-ec11-43d3-84b4-1b3696caf977","Type":"ContainerStarted","Data":"2f6b327481ce1c5b46128acf3fd6bd719d49f8e1449315420c62efb062c67050"} Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.722573 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d"] Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.723652 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.728236 5018 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nwcm7" Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.736570 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d"] Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.893869 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrmjt\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-kube-api-access-zrmjt\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.894084 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.996063 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrmjt\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-kube-api-access-zrmjt\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:45 crc kubenswrapper[5018]: I1014 07:04:45.996144 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:46 crc kubenswrapper[5018]: I1014 07:04:46.020966 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:46 crc kubenswrapper[5018]: I1014 07:04:46.027455 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrmjt\" (UniqueName: \"kubernetes.io/projected/07f3eb84-992a-499b-8916-a349d9defeec-kube-api-access-zrmjt\") pod \"cert-manager-cainjector-7d9f95dbf-29g6d\" (UID: \"07f3eb84-992a-499b-8916-a349d9defeec\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:46 crc kubenswrapper[5018]: I1014 07:04:46.054613 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" Oct 14 07:04:46 crc kubenswrapper[5018]: I1014 07:04:46.330717 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d"] Oct 14 07:04:46 crc kubenswrapper[5018]: W1014 07:04:46.339590 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07f3eb84_992a_499b_8916_a349d9defeec.slice/crio-8ad6427f08e2236ee794de5bd8a230785dc399ba1476e30cbb578f90e5ae220a WatchSource:0}: Error finding container 8ad6427f08e2236ee794de5bd8a230785dc399ba1476e30cbb578f90e5ae220a: Status 404 returned error can't find the container with id 8ad6427f08e2236ee794de5bd8a230785dc399ba1476e30cbb578f90e5ae220a Oct 14 07:04:46 crc kubenswrapper[5018]: I1014 07:04:46.423251 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" event={"ID":"07f3eb84-992a-499b-8916-a349d9defeec","Type":"ContainerStarted","Data":"8ad6427f08e2236ee794de5bd8a230785dc399ba1476e30cbb578f90e5ae220a"} Oct 14 07:04:54 crc kubenswrapper[5018]: I1014 07:04:54.491056 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-sb596" event={"ID":"e7159c40-ec11-43d3-84b4-1b3696caf977","Type":"ContainerStarted","Data":"52ca0745c433c731be619088b23e18f61b9f0ff24dceb7396a248efe968c70dc"} Oct 14 07:04:54 crc kubenswrapper[5018]: I1014 07:04:54.491958 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:04:54 crc kubenswrapper[5018]: I1014 07:04:54.492957 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" event={"ID":"07f3eb84-992a-499b-8916-a349d9defeec","Type":"ContainerStarted","Data":"47331d6ebe72a6e1367d3a6c70379666434b6760ea4a3edb6153df887b87e38b"} Oct 14 07:04:54 crc kubenswrapper[5018]: I1014 07:04:54.522006 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-sb596" podStartSLOduration=1.858730682 podStartE2EDuration="11.521981702s" podCreationTimestamp="2025-10-14 07:04:43 +0000 UTC" firstStartedPulling="2025-10-14 07:04:43.939445373 +0000 UTC m=+900.523491990" lastFinishedPulling="2025-10-14 07:04:53.602696393 +0000 UTC m=+910.186743010" observedRunningTime="2025-10-14 07:04:54.517336833 +0000 UTC m=+911.101383480" watchObservedRunningTime="2025-10-14 07:04:54.521981702 +0000 UTC m=+911.106028359" Oct 14 07:04:54 crc kubenswrapper[5018]: I1014 07:04:54.536179 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-29g6d" podStartSLOduration=2.252929268 podStartE2EDuration="9.536156888s" podCreationTimestamp="2025-10-14 07:04:45 +0000 UTC" firstStartedPulling="2025-10-14 07:04:46.342092843 +0000 UTC m=+902.926139470" lastFinishedPulling="2025-10-14 07:04:53.625320463 +0000 UTC m=+910.209367090" observedRunningTime="2025-10-14 07:04:54.532824235 +0000 UTC m=+911.116870872" watchObservedRunningTime="2025-10-14 07:04:54.536156888 +0000 UTC m=+911.120203525" Oct 14 07:04:58 crc kubenswrapper[5018]: I1014 07:04:58.483268 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-sb596" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.447566 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-czf65"] Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.449468 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.466434 5018 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8dd8j" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.469543 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-czf65"] Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.510532 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whxs6\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-kube-api-access-whxs6\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.510686 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.611853 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.612196 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whxs6\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-kube-api-access-whxs6\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.642608 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.642889 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whxs6\" (UniqueName: \"kubernetes.io/projected/3212416e-b631-4294-83fc-75b350569b44-kube-api-access-whxs6\") pod \"cert-manager-7d4cc89fcb-czf65\" (UID: \"3212416e-b631-4294-83fc-75b350569b44\") " pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:01 crc kubenswrapper[5018]: I1014 07:05:01.786528 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-czf65" Oct 14 07:05:02 crc kubenswrapper[5018]: I1014 07:05:02.295012 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-czf65"] Oct 14 07:05:02 crc kubenswrapper[5018]: I1014 07:05:02.463693 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:05:02 crc kubenswrapper[5018]: I1014 07:05:02.463779 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:05:02 crc kubenswrapper[5018]: I1014 07:05:02.548832 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-czf65" event={"ID":"3212416e-b631-4294-83fc-75b350569b44","Type":"ContainerStarted","Data":"f7a923475a3da28f2db9b783cf3680b77cd54cecb77f0d9d8f6ba01215d02d7a"} Oct 14 07:05:02 crc kubenswrapper[5018]: I1014 07:05:02.549378 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-czf65" event={"ID":"3212416e-b631-4294-83fc-75b350569b44","Type":"ContainerStarted","Data":"0b139864e9d67932e788e2fbff4052571fc3032258b995b27f970693ade4b3ec"} Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.410937 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-czf65" podStartSLOduration=11.410911227 podStartE2EDuration="11.410911227s" podCreationTimestamp="2025-10-14 07:05:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:05:02.566229262 +0000 UTC m=+919.150275879" watchObservedRunningTime="2025-10-14 07:05:12.410911227 +0000 UTC m=+928.994957894" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.417719 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.418940 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.421190 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.423341 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-wc9jt" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.423819 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.441425 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.578728 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gts2n\" (UniqueName: \"kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n\") pod \"openstack-operator-index-mgl25\" (UID: \"763c0e0f-dab6-439d-ae57-f15fa6840715\") " pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.679913 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gts2n\" (UniqueName: \"kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n\") pod \"openstack-operator-index-mgl25\" (UID: \"763c0e0f-dab6-439d-ae57-f15fa6840715\") " pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.709384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gts2n\" (UniqueName: \"kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n\") pod \"openstack-operator-index-mgl25\" (UID: \"763c0e0f-dab6-439d-ae57-f15fa6840715\") " pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:12 crc kubenswrapper[5018]: I1014 07:05:12.742805 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:13 crc kubenswrapper[5018]: I1014 07:05:13.173951 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:13 crc kubenswrapper[5018]: I1014 07:05:13.627350 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mgl25" event={"ID":"763c0e0f-dab6-439d-ae57-f15fa6840715","Type":"ContainerStarted","Data":"9466101b92693aad4edfcdf99f47954af4436c0d901951c346d29e496bad8a9d"} Oct 14 07:05:14 crc kubenswrapper[5018]: I1014 07:05:14.647136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mgl25" event={"ID":"763c0e0f-dab6-439d-ae57-f15fa6840715","Type":"ContainerStarted","Data":"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55"} Oct 14 07:05:14 crc kubenswrapper[5018]: I1014 07:05:14.667250 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mgl25" podStartSLOduration=1.670150861 podStartE2EDuration="2.667233715s" podCreationTimestamp="2025-10-14 07:05:12 +0000 UTC" firstStartedPulling="2025-10-14 07:05:13.18884633 +0000 UTC m=+929.772892997" lastFinishedPulling="2025-10-14 07:05:14.185929194 +0000 UTC m=+930.769975851" observedRunningTime="2025-10-14 07:05:14.665353373 +0000 UTC m=+931.249400000" watchObservedRunningTime="2025-10-14 07:05:14.667233715 +0000 UTC m=+931.251280342" Oct 14 07:05:15 crc kubenswrapper[5018]: I1014 07:05:15.178552 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:15 crc kubenswrapper[5018]: I1014 07:05:15.791890 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vlqj5"] Oct 14 07:05:15 crc kubenswrapper[5018]: I1014 07:05:15.793074 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:15 crc kubenswrapper[5018]: I1014 07:05:15.806558 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vlqj5"] Oct 14 07:05:15 crc kubenswrapper[5018]: I1014 07:05:15.926050 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48qsp\" (UniqueName: \"kubernetes.io/projected/72081dcc-e05e-4d65-bc3b-cf6a398a66d2-kube-api-access-48qsp\") pod \"openstack-operator-index-vlqj5\" (UID: \"72081dcc-e05e-4d65-bc3b-cf6a398a66d2\") " pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.027956 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48qsp\" (UniqueName: \"kubernetes.io/projected/72081dcc-e05e-4d65-bc3b-cf6a398a66d2-kube-api-access-48qsp\") pod \"openstack-operator-index-vlqj5\" (UID: \"72081dcc-e05e-4d65-bc3b-cf6a398a66d2\") " pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.062235 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48qsp\" (UniqueName: \"kubernetes.io/projected/72081dcc-e05e-4d65-bc3b-cf6a398a66d2-kube-api-access-48qsp\") pod \"openstack-operator-index-vlqj5\" (UID: \"72081dcc-e05e-4d65-bc3b-cf6a398a66d2\") " pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.120320 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.598979 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vlqj5"] Oct 14 07:05:16 crc kubenswrapper[5018]: W1014 07:05:16.608370 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72081dcc_e05e_4d65_bc3b_cf6a398a66d2.slice/crio-8f3444f85f059639a7d1a8ca140abd816da1f25fec0dea9b8ccc251ceb55ebb3 WatchSource:0}: Error finding container 8f3444f85f059639a7d1a8ca140abd816da1f25fec0dea9b8ccc251ceb55ebb3: Status 404 returned error can't find the container with id 8f3444f85f059639a7d1a8ca140abd816da1f25fec0dea9b8ccc251ceb55ebb3 Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.664014 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vlqj5" event={"ID":"72081dcc-e05e-4d65-bc3b-cf6a398a66d2","Type":"ContainerStarted","Data":"8f3444f85f059639a7d1a8ca140abd816da1f25fec0dea9b8ccc251ceb55ebb3"} Oct 14 07:05:16 crc kubenswrapper[5018]: I1014 07:05:16.664133 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-mgl25" podUID="763c0e0f-dab6-439d-ae57-f15fa6840715" containerName="registry-server" containerID="cri-o://be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55" gracePeriod=2 Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.161604 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.345510 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gts2n\" (UniqueName: \"kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n\") pod \"763c0e0f-dab6-439d-ae57-f15fa6840715\" (UID: \"763c0e0f-dab6-439d-ae57-f15fa6840715\") " Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.351957 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n" (OuterVolumeSpecName: "kube-api-access-gts2n") pod "763c0e0f-dab6-439d-ae57-f15fa6840715" (UID: "763c0e0f-dab6-439d-ae57-f15fa6840715"). InnerVolumeSpecName "kube-api-access-gts2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.448715 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gts2n\" (UniqueName: \"kubernetes.io/projected/763c0e0f-dab6-439d-ae57-f15fa6840715-kube-api-access-gts2n\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.674484 5018 generic.go:334] "Generic (PLEG): container finished" podID="763c0e0f-dab6-439d-ae57-f15fa6840715" containerID="be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55" exitCode=0 Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.674548 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mgl25" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.674941 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mgl25" event={"ID":"763c0e0f-dab6-439d-ae57-f15fa6840715","Type":"ContainerDied","Data":"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55"} Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.675035 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mgl25" event={"ID":"763c0e0f-dab6-439d-ae57-f15fa6840715","Type":"ContainerDied","Data":"9466101b92693aad4edfcdf99f47954af4436c0d901951c346d29e496bad8a9d"} Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.675097 5018 scope.go:117] "RemoveContainer" containerID="be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.676692 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vlqj5" event={"ID":"72081dcc-e05e-4d65-bc3b-cf6a398a66d2","Type":"ContainerStarted","Data":"b6ac362a6ae56ae2d6bf7f87b194c2e98f2b696342638831b32ac29dd0fe1f7c"} Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.704764 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vlqj5" podStartSLOduration=2.270857739 podStartE2EDuration="2.704736748s" podCreationTimestamp="2025-10-14 07:05:15 +0000 UTC" firstStartedPulling="2025-10-14 07:05:16.610385542 +0000 UTC m=+933.194432189" lastFinishedPulling="2025-10-14 07:05:17.044264561 +0000 UTC m=+933.628311198" observedRunningTime="2025-10-14 07:05:17.703350929 +0000 UTC m=+934.287397586" watchObservedRunningTime="2025-10-14 07:05:17.704736748 +0000 UTC m=+934.288783415" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.712896 5018 scope.go:117] "RemoveContainer" containerID="be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55" Oct 14 07:05:17 crc kubenswrapper[5018]: E1014 07:05:17.713776 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55\": container with ID starting with be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55 not found: ID does not exist" containerID="be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.713876 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55"} err="failed to get container status \"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55\": rpc error: code = NotFound desc = could not find container \"be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55\": container with ID starting with be6ef7e2190d7b05dbfc8734c5105175b0863638caffcee5e4bb7420e7490f55 not found: ID does not exist" Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.735479 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:17 crc kubenswrapper[5018]: I1014 07:05:17.741500 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-mgl25"] Oct 14 07:05:18 crc kubenswrapper[5018]: I1014 07:05:18.619649 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="763c0e0f-dab6-439d-ae57-f15fa6840715" path="/var/lib/kubelet/pods/763c0e0f-dab6-439d-ae57-f15fa6840715/volumes" Oct 14 07:05:26 crc kubenswrapper[5018]: I1014 07:05:26.121338 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:26 crc kubenswrapper[5018]: I1014 07:05:26.122104 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:26 crc kubenswrapper[5018]: I1014 07:05:26.157186 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:26 crc kubenswrapper[5018]: I1014 07:05:26.801720 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vlqj5" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.463038 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.463530 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.495388 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg"] Oct 14 07:05:32 crc kubenswrapper[5018]: E1014 07:05:32.495668 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763c0e0f-dab6-439d-ae57-f15fa6840715" containerName="registry-server" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.495682 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="763c0e0f-dab6-439d-ae57-f15fa6840715" containerName="registry-server" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.495825 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="763c0e0f-dab6-439d-ae57-f15fa6840715" containerName="registry-server" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.496820 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.498654 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqxb" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.510809 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg"] Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.680032 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.680222 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.680559 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sflcw\" (UniqueName: \"kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.781946 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.782450 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sflcw\" (UniqueName: \"kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.782738 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.782759 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.784035 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.812057 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sflcw\" (UniqueName: \"kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:32 crc kubenswrapper[5018]: I1014 07:05:32.828106 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:33 crc kubenswrapper[5018]: I1014 07:05:33.133189 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg"] Oct 14 07:05:33 crc kubenswrapper[5018]: W1014 07:05:33.135474 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ba36249_21fb_410f_895a_19cccb4e5fb1.slice/crio-a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7 WatchSource:0}: Error finding container a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7: Status 404 returned error can't find the container with id a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7 Oct 14 07:05:33 crc kubenswrapper[5018]: I1014 07:05:33.854917 5018 generic.go:334] "Generic (PLEG): container finished" podID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerID="67a3aab754debcd041dc6aa906b9eb5fae0b9d24571686685784a0b0df354470" exitCode=0 Oct 14 07:05:33 crc kubenswrapper[5018]: I1014 07:05:33.855026 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" event={"ID":"7ba36249-21fb-410f-895a-19cccb4e5fb1","Type":"ContainerDied","Data":"67a3aab754debcd041dc6aa906b9eb5fae0b9d24571686685784a0b0df354470"} Oct 14 07:05:33 crc kubenswrapper[5018]: I1014 07:05:33.855307 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" event={"ID":"7ba36249-21fb-410f-895a-19cccb4e5fb1","Type":"ContainerStarted","Data":"a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7"} Oct 14 07:05:34 crc kubenswrapper[5018]: I1014 07:05:34.862847 5018 generic.go:334] "Generic (PLEG): container finished" podID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerID="a97bb9e31ae56cf1575fdb49516b8650c2e045ac262a43964918c54adc72e2f8" exitCode=0 Oct 14 07:05:34 crc kubenswrapper[5018]: I1014 07:05:34.862904 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" event={"ID":"7ba36249-21fb-410f-895a-19cccb4e5fb1","Type":"ContainerDied","Data":"a97bb9e31ae56cf1575fdb49516b8650c2e045ac262a43964918c54adc72e2f8"} Oct 14 07:05:35 crc kubenswrapper[5018]: I1014 07:05:35.873218 5018 generic.go:334] "Generic (PLEG): container finished" podID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerID="8a037358cc21326420e37ec0e7488e9447dcd0755372facdbd077d08200884ce" exitCode=0 Oct 14 07:05:35 crc kubenswrapper[5018]: I1014 07:05:35.873280 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" event={"ID":"7ba36249-21fb-410f-895a-19cccb4e5fb1","Type":"ContainerDied","Data":"8a037358cc21326420e37ec0e7488e9447dcd0755372facdbd077d08200884ce"} Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.202137 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.350995 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sflcw\" (UniqueName: \"kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw\") pod \"7ba36249-21fb-410f-895a-19cccb4e5fb1\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.351075 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util\") pod \"7ba36249-21fb-410f-895a-19cccb4e5fb1\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.351228 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle\") pod \"7ba36249-21fb-410f-895a-19cccb4e5fb1\" (UID: \"7ba36249-21fb-410f-895a-19cccb4e5fb1\") " Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.352910 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle" (OuterVolumeSpecName: "bundle") pod "7ba36249-21fb-410f-895a-19cccb4e5fb1" (UID: "7ba36249-21fb-410f-895a-19cccb4e5fb1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.357143 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw" (OuterVolumeSpecName: "kube-api-access-sflcw") pod "7ba36249-21fb-410f-895a-19cccb4e5fb1" (UID: "7ba36249-21fb-410f-895a-19cccb4e5fb1"). InnerVolumeSpecName "kube-api-access-sflcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.380489 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util" (OuterVolumeSpecName: "util") pod "7ba36249-21fb-410f-895a-19cccb4e5fb1" (UID: "7ba36249-21fb-410f-895a-19cccb4e5fb1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.453410 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sflcw\" (UniqueName: \"kubernetes.io/projected/7ba36249-21fb-410f-895a-19cccb4e5fb1-kube-api-access-sflcw\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.453459 5018 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.453480 5018 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ba36249-21fb-410f-895a-19cccb4e5fb1-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.891969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" event={"ID":"7ba36249-21fb-410f-895a-19cccb4e5fb1","Type":"ContainerDied","Data":"a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7"} Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.892296 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a826dda5a584721e51eff88420896a2c9c18bd239f7083dbc51e449b7604d4b7" Oct 14 07:05:37 crc kubenswrapper[5018]: I1014 07:05:37.892078 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.132929 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw"] Oct 14 07:05:40 crc kubenswrapper[5018]: E1014 07:05:40.134209 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="util" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.134312 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="util" Oct 14 07:05:40 crc kubenswrapper[5018]: E1014 07:05:40.134381 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="extract" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.134433 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="extract" Oct 14 07:05:40 crc kubenswrapper[5018]: E1014 07:05:40.134492 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="pull" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.134544 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="pull" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.134711 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba36249-21fb-410f-895a-19cccb4e5fb1" containerName="extract" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.135305 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.137295 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xfd6z" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.157210 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw"] Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.290651 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw7qt\" (UniqueName: \"kubernetes.io/projected/50db8b96-a5ef-46d9-ba42-4f41d61ff536-kube-api-access-vw7qt\") pod \"openstack-operator-controller-operator-64895cd698-5vxbw\" (UID: \"50db8b96-a5ef-46d9-ba42-4f41d61ff536\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.392180 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw7qt\" (UniqueName: \"kubernetes.io/projected/50db8b96-a5ef-46d9-ba42-4f41d61ff536-kube-api-access-vw7qt\") pod \"openstack-operator-controller-operator-64895cd698-5vxbw\" (UID: \"50db8b96-a5ef-46d9-ba42-4f41d61ff536\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.422900 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw7qt\" (UniqueName: \"kubernetes.io/projected/50db8b96-a5ef-46d9-ba42-4f41d61ff536-kube-api-access-vw7qt\") pod \"openstack-operator-controller-operator-64895cd698-5vxbw\" (UID: \"50db8b96-a5ef-46d9-ba42-4f41d61ff536\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.450525 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.874495 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw"] Oct 14 07:05:40 crc kubenswrapper[5018]: I1014 07:05:40.911451 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" event={"ID":"50db8b96-a5ef-46d9-ba42-4f41d61ff536","Type":"ContainerStarted","Data":"06d9c826753026c884710c547dd49391cd532823e2ca0cc77907148e2029e0d3"} Oct 14 07:05:45 crc kubenswrapper[5018]: I1014 07:05:45.946361 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" event={"ID":"50db8b96-a5ef-46d9-ba42-4f41d61ff536","Type":"ContainerStarted","Data":"236abd2ccc7f4a54df7c0fd932923e020f3bcd4391f3de4ad23af775d8640361"} Oct 14 07:05:47 crc kubenswrapper[5018]: I1014 07:05:47.961594 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" event={"ID":"50db8b96-a5ef-46d9-ba42-4f41d61ff536","Type":"ContainerStarted","Data":"b9aa8e0e795817d14e34f1864be40510da0901add893f96e239b9eca00e340a8"} Oct 14 07:05:47 crc kubenswrapper[5018]: I1014 07:05:47.962093 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:05:48 crc kubenswrapper[5018]: I1014 07:05:48.020697 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" podStartSLOduration=1.6352814690000002 podStartE2EDuration="8.02066393s" podCreationTimestamp="2025-10-14 07:05:40 +0000 UTC" firstStartedPulling="2025-10-14 07:05:40.887254229 +0000 UTC m=+957.471300856" lastFinishedPulling="2025-10-14 07:05:47.27263667 +0000 UTC m=+963.856683317" observedRunningTime="2025-10-14 07:05:48.01243046 +0000 UTC m=+964.596477137" watchObservedRunningTime="2025-10-14 07:05:48.02066393 +0000 UTC m=+964.604710597" Oct 14 07:05:50 crc kubenswrapper[5018]: I1014 07:05:50.453830 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-5vxbw" Oct 14 07:06:02 crc kubenswrapper[5018]: I1014 07:06:02.464001 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:06:02 crc kubenswrapper[5018]: I1014 07:06:02.464266 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:06:02 crc kubenswrapper[5018]: I1014 07:06:02.464314 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:06:02 crc kubenswrapper[5018]: I1014 07:06:02.464922 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:06:02 crc kubenswrapper[5018]: I1014 07:06:02.464987 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986" gracePeriod=600 Oct 14 07:06:03 crc kubenswrapper[5018]: I1014 07:06:03.059926 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986" exitCode=0 Oct 14 07:06:03 crc kubenswrapper[5018]: I1014 07:06:03.059976 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986"} Oct 14 07:06:03 crc kubenswrapper[5018]: I1014 07:06:03.060320 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd"} Oct 14 07:06:03 crc kubenswrapper[5018]: I1014 07:06:03.060347 5018 scope.go:117] "RemoveContainer" containerID="7b75b0e42e5fac12a50bfa0072bd12a991ad51da977009cdb7c57787989e4345" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.467374 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.468944 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.483581 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qsqsn" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.491563 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.492461 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.498018 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qhmtl" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.498205 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.505549 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.506490 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.507477 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62fz\" (UniqueName: \"kubernetes.io/projected/a0e4cf91-8413-4a0f-b5ec-06f8becca6a7-kube-api-access-m62fz\") pod \"barbican-operator-controller-manager-658bdf4b74-swhnp\" (UID: \"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.517785 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.518099 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zhmjt" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.526315 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.545686 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.546732 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.550782 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-df2fs" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.559565 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.560430 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.562552 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sjfvf" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.580491 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.581774 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.590021 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ftl6n" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.602267 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.608901 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgjlc\" (UniqueName: \"kubernetes.io/projected/0e01bea2-81b7-4b83-881c-c40f3a8bd306-kube-api-access-xgjlc\") pod \"cinder-operator-controller-manager-7b7fb68549-vwpjg\" (UID: \"0e01bea2-81b7-4b83-881c-c40f3a8bd306\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.608943 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsj57\" (UniqueName: \"kubernetes.io/projected/d94d1327-34a3-45fe-963c-b80cc8931811-kube-api-access-tsj57\") pod \"glance-operator-controller-manager-84b9b84486-gnt2q\" (UID: \"d94d1327-34a3-45fe-963c-b80cc8931811\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.608978 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9qp\" (UniqueName: \"kubernetes.io/projected/284f0e2f-1e97-4201-8313-0c604eb245c7-kube-api-access-4x9qp\") pod \"heat-operator-controller-manager-858f76bbdd-c6q56\" (UID: \"284f0e2f-1e97-4201-8313-0c604eb245c7\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.609002 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46hxd\" (UniqueName: \"kubernetes.io/projected/88f2a077-948e-4245-834d-9f943670b40f-kube-api-access-46hxd\") pod \"horizon-operator-controller-manager-7ffbcb7588-b9n5t\" (UID: \"88f2a077-948e-4245-834d-9f943670b40f\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.609027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqbxk\" (UniqueName: \"kubernetes.io/projected/a81fad02-cd39-487a-ba7d-d29757b26848-kube-api-access-lqbxk\") pod \"designate-operator-controller-manager-85d5d9dd78-4fgw9\" (UID: \"a81fad02-cd39-487a-ba7d-d29757b26848\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.609078 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62fz\" (UniqueName: \"kubernetes.io/projected/a0e4cf91-8413-4a0f-b5ec-06f8becca6a7-kube-api-access-m62fz\") pod \"barbican-operator-controller-manager-658bdf4b74-swhnp\" (UID: \"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.612678 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.621991 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.622973 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.634794 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.635118 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-zkwhs" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.654740 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.656273 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62fz\" (UniqueName: \"kubernetes.io/projected/a0e4cf91-8413-4a0f-b5ec-06f8becca6a7-kube-api-access-m62fz\") pod \"barbican-operator-controller-manager-658bdf4b74-swhnp\" (UID: \"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.688775 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710501 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqbxk\" (UniqueName: \"kubernetes.io/projected/a81fad02-cd39-487a-ba7d-d29757b26848-kube-api-access-lqbxk\") pod \"designate-operator-controller-manager-85d5d9dd78-4fgw9\" (UID: \"a81fad02-cd39-487a-ba7d-d29757b26848\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710594 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlg87\" (UniqueName: \"kubernetes.io/projected/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-kube-api-access-qlg87\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710656 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgjlc\" (UniqueName: \"kubernetes.io/projected/0e01bea2-81b7-4b83-881c-c40f3a8bd306-kube-api-access-xgjlc\") pod \"cinder-operator-controller-manager-7b7fb68549-vwpjg\" (UID: \"0e01bea2-81b7-4b83-881c-c40f3a8bd306\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710678 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsj57\" (UniqueName: \"kubernetes.io/projected/d94d1327-34a3-45fe-963c-b80cc8931811-kube-api-access-tsj57\") pod \"glance-operator-controller-manager-84b9b84486-gnt2q\" (UID: \"d94d1327-34a3-45fe-963c-b80cc8931811\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710713 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9qp\" (UniqueName: \"kubernetes.io/projected/284f0e2f-1e97-4201-8313-0c604eb245c7-kube-api-access-4x9qp\") pod \"heat-operator-controller-manager-858f76bbdd-c6q56\" (UID: \"284f0e2f-1e97-4201-8313-0c604eb245c7\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710738 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.710761 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46hxd\" (UniqueName: \"kubernetes.io/projected/88f2a077-948e-4245-834d-9f943670b40f-kube-api-access-46hxd\") pod \"horizon-operator-controller-manager-7ffbcb7588-b9n5t\" (UID: \"88f2a077-948e-4245-834d-9f943670b40f\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.735451 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.738404 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.742826 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqbxk\" (UniqueName: \"kubernetes.io/projected/a81fad02-cd39-487a-ba7d-d29757b26848-kube-api-access-lqbxk\") pod \"designate-operator-controller-manager-85d5d9dd78-4fgw9\" (UID: \"a81fad02-cd39-487a-ba7d-d29757b26848\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.746552 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.747652 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.749862 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-sqnfr" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.751170 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mqg68" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.751852 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46hxd\" (UniqueName: \"kubernetes.io/projected/88f2a077-948e-4245-834d-9f943670b40f-kube-api-access-46hxd\") pod \"horizon-operator-controller-manager-7ffbcb7588-b9n5t\" (UID: \"88f2a077-948e-4245-834d-9f943670b40f\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.758175 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgjlc\" (UniqueName: \"kubernetes.io/projected/0e01bea2-81b7-4b83-881c-c40f3a8bd306-kube-api-access-xgjlc\") pod \"cinder-operator-controller-manager-7b7fb68549-vwpjg\" (UID: \"0e01bea2-81b7-4b83-881c-c40f3a8bd306\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.759374 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.777236 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9qp\" (UniqueName: \"kubernetes.io/projected/284f0e2f-1e97-4201-8313-0c604eb245c7-kube-api-access-4x9qp\") pod \"heat-operator-controller-manager-858f76bbdd-c6q56\" (UID: \"284f0e2f-1e97-4201-8313-0c604eb245c7\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.778416 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsj57\" (UniqueName: \"kubernetes.io/projected/d94d1327-34a3-45fe-963c-b80cc8931811-kube-api-access-tsj57\") pod \"glance-operator-controller-manager-84b9b84486-gnt2q\" (UID: \"d94d1327-34a3-45fe-963c-b80cc8931811\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.787884 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.788191 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.799100 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.799999 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.808991 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-kghk6" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.811851 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nxrz\" (UniqueName: \"kubernetes.io/projected/4afa0a87-9ea4-4d10-857d-8c046132dbd3-kube-api-access-9nxrz\") pod \"keystone-operator-controller-manager-55b6b7c7b8-wrxs4\" (UID: \"4afa0a87-9ea4-4d10-857d-8c046132dbd3\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.812016 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlg87\" (UniqueName: \"kubernetes.io/projected/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-kube-api-access-qlg87\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.812101 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5lwd\" (UniqueName: \"kubernetes.io/projected/3195f44a-f801-4f06-ae14-d83a680343a7-kube-api-access-k5lwd\") pod \"manila-operator-controller-manager-5f67fbc655-v2ppk\" (UID: \"3195f44a-f801-4f06-ae14-d83a680343a7\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.812222 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.812301 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdr72\" (UniqueName: \"kubernetes.io/projected/a3965916-7dc3-4bfc-ae94-c57ef9b4b43c-kube-api-access-cdr72\") pod \"ironic-operator-controller-manager-9c5c78d49-jrnfz\" (UID: \"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:07 crc kubenswrapper[5018]: E1014 07:06:07.812571 5018 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 14 07:06:07 crc kubenswrapper[5018]: E1014 07:06:07.812680 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert podName:dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff nodeName:}" failed. No retries permitted until 2025-10-14 07:06:08.312660624 +0000 UTC m=+984.896707241 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert") pod "infra-operator-controller-manager-656bcbd775-tfcm7" (UID: "dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff") : secret "infra-operator-webhook-server-cert" not found Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.815649 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.816856 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.818720 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.818996 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.824294 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.825154 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.832818 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.832865 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.842891 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-dhc98"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.843712 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-dhc98"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.843781 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.844747 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.847534 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-m4fdn" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.847723 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-s6294" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.847858 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9764b" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.861686 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.862807 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.865466 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.865751 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rrh6q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.869911 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.872981 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.874051 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.875474 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-h5hqx" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.880459 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.882700 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.885003 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlg87\" (UniqueName: \"kubernetes.io/projected/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-kube-api-access-qlg87\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.885308 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.887204 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-d85js" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.887360 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.896823 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.911754 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.912032 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.916452 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-spv54" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.917022 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.920687 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.920951 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44m8\" (UniqueName: \"kubernetes.io/projected/df3020fc-9cba-409f-8355-c76ddfb47db2-kube-api-access-n44m8\") pod \"ovn-operator-controller-manager-79df5fb58c-wltkk\" (UID: \"df3020fc-9cba-409f-8355-c76ddfb47db2\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921023 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsb6c\" (UniqueName: \"kubernetes.io/projected/6bf6a59f-81bd-4003-a387-32c703648d21-kube-api-access-zsb6c\") pod \"mariadb-operator-controller-manager-f9fb45f8f-92pt5\" (UID: \"6bf6a59f-81bd-4003-a387-32c703648d21\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921047 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6lmv\" (UniqueName: \"kubernetes.io/projected/fb295ac9-43c8-4482-b557-9ac8aff33171-kube-api-access-t6lmv\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921073 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdr72\" (UniqueName: \"kubernetes.io/projected/a3965916-7dc3-4bfc-ae94-c57ef9b4b43c-kube-api-access-cdr72\") pod \"ironic-operator-controller-manager-9c5c78d49-jrnfz\" (UID: \"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921110 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hg2s\" (UniqueName: \"kubernetes.io/projected/15fc314b-1904-488c-829c-9c5c0274ff20-kube-api-access-2hg2s\") pod \"nova-operator-controller-manager-5df598886f-dhc98\" (UID: \"15fc314b-1904-488c-829c-9c5c0274ff20\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921133 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nxrz\" (UniqueName: \"kubernetes.io/projected/4afa0a87-9ea4-4d10-857d-8c046132dbd3-kube-api-access-9nxrz\") pod \"keystone-operator-controller-manager-55b6b7c7b8-wrxs4\" (UID: \"4afa0a87-9ea4-4d10-857d-8c046132dbd3\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921159 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921182 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzpqs\" (UniqueName: \"kubernetes.io/projected/3c93e5a0-bd99-4348-8562-6584ca640fc8-kube-api-access-dzpqs\") pod \"neutron-operator-controller-manager-79d585cb66-rfcrs\" (UID: \"3c93e5a0-bd99-4348-8562-6584ca640fc8\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921215 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5m2m\" (UniqueName: \"kubernetes.io/projected/5fea696b-3ec3-47c7-bc33-212456367942-kube-api-access-k5m2m\") pod \"octavia-operator-controller-manager-69fdcfc5f5-mc6cb\" (UID: \"5fea696b-3ec3-47c7-bc33-212456367942\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.921239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5lwd\" (UniqueName: \"kubernetes.io/projected/3195f44a-f801-4f06-ae14-d83a680343a7-kube-api-access-k5lwd\") pod \"manila-operator-controller-manager-5f67fbc655-v2ppk\" (UID: \"3195f44a-f801-4f06-ae14-d83a680343a7\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.927082 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.928337 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.930841 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-g9hzh" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.933512 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.944765 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdr72\" (UniqueName: \"kubernetes.io/projected/a3965916-7dc3-4bfc-ae94-c57ef9b4b43c-kube-api-access-cdr72\") pod \"ironic-operator-controller-manager-9c5c78d49-jrnfz\" (UID: \"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.946456 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5lwd\" (UniqueName: \"kubernetes.io/projected/3195f44a-f801-4f06-ae14-d83a680343a7-kube-api-access-k5lwd\") pod \"manila-operator-controller-manager-5f67fbc655-v2ppk\" (UID: \"3195f44a-f801-4f06-ae14-d83a680343a7\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.948339 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.949227 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nxrz\" (UniqueName: \"kubernetes.io/projected/4afa0a87-9ea4-4d10-857d-8c046132dbd3-kube-api-access-9nxrz\") pod \"keystone-operator-controller-manager-55b6b7c7b8-wrxs4\" (UID: \"4afa0a87-9ea4-4d10-857d-8c046132dbd3\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.967970 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn"] Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.969992 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.971758 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-sh9bf" Oct 14 07:06:07 crc kubenswrapper[5018]: I1014 07:06:07.973467 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024044 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwmxn\" (UniqueName: \"kubernetes.io/projected/e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f-kube-api-access-xwmxn\") pod \"placement-operator-controller-manager-68b6c87b68-vx59l\" (UID: \"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024078 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjgnp\" (UniqueName: \"kubernetes.io/projected/dba339e1-27ec-4b88-aa67-b1d72aced19a-kube-api-access-fjgnp\") pod \"telemetry-operator-controller-manager-67cfc6749b-pgbzn\" (UID: \"dba339e1-27ec-4b88-aa67-b1d72aced19a\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024117 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hg2s\" (UniqueName: \"kubernetes.io/projected/15fc314b-1904-488c-829c-9c5c0274ff20-kube-api-access-2hg2s\") pod \"nova-operator-controller-manager-5df598886f-dhc98\" (UID: \"15fc314b-1904-488c-829c-9c5c0274ff20\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024150 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024173 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzpqs\" (UniqueName: \"kubernetes.io/projected/3c93e5a0-bd99-4348-8562-6584ca640fc8-kube-api-access-dzpqs\") pod \"neutron-operator-controller-manager-79d585cb66-rfcrs\" (UID: \"3c93e5a0-bd99-4348-8562-6584ca640fc8\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5m2m\" (UniqueName: \"kubernetes.io/projected/5fea696b-3ec3-47c7-bc33-212456367942-kube-api-access-k5m2m\") pod \"octavia-operator-controller-manager-69fdcfc5f5-mc6cb\" (UID: \"5fea696b-3ec3-47c7-bc33-212456367942\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024249 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44m8\" (UniqueName: \"kubernetes.io/projected/df3020fc-9cba-409f-8355-c76ddfb47db2-kube-api-access-n44m8\") pod \"ovn-operator-controller-manager-79df5fb58c-wltkk\" (UID: \"df3020fc-9cba-409f-8355-c76ddfb47db2\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024289 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsb6c\" (UniqueName: \"kubernetes.io/projected/6bf6a59f-81bd-4003-a387-32c703648d21-kube-api-access-zsb6c\") pod \"mariadb-operator-controller-manager-f9fb45f8f-92pt5\" (UID: \"6bf6a59f-81bd-4003-a387-32c703648d21\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024307 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6lmv\" (UniqueName: \"kubernetes.io/projected/fb295ac9-43c8-4482-b557-9ac8aff33171-kube-api-access-t6lmv\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.024329 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdscf\" (UniqueName: \"kubernetes.io/projected/fdb571a6-0d71-4cd9-885d-798b4489b816-kube-api-access-vdscf\") pod \"swift-operator-controller-manager-db6d7f97b-s48sn\" (UID: \"fdb571a6-0d71-4cd9-885d-798b4489b816\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.024891 5018 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.024938 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert podName:fb295ac9-43c8-4482-b557-9ac8aff33171 nodeName:}" failed. No retries permitted until 2025-10-14 07:06:08.524924783 +0000 UTC m=+985.108971410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert") pod "openstack-baremetal-operator-controller-manager-55b7d448487bng8" (UID: "fb295ac9-43c8-4482-b557-9ac8aff33171") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.032403 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.034395 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.039392 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-56fkc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.079325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsb6c\" (UniqueName: \"kubernetes.io/projected/6bf6a59f-81bd-4003-a387-32c703648d21-kube-api-access-zsb6c\") pod \"mariadb-operator-controller-manager-f9fb45f8f-92pt5\" (UID: \"6bf6a59f-81bd-4003-a387-32c703648d21\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.081949 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44m8\" (UniqueName: \"kubernetes.io/projected/df3020fc-9cba-409f-8355-c76ddfb47db2-kube-api-access-n44m8\") pod \"ovn-operator-controller-manager-79df5fb58c-wltkk\" (UID: \"df3020fc-9cba-409f-8355-c76ddfb47db2\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.092002 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.098894 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzpqs\" (UniqueName: \"kubernetes.io/projected/3c93e5a0-bd99-4348-8562-6584ca640fc8-kube-api-access-dzpqs\") pod \"neutron-operator-controller-manager-79d585cb66-rfcrs\" (UID: \"3c93e5a0-bd99-4348-8562-6584ca640fc8\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.102058 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5m2m\" (UniqueName: \"kubernetes.io/projected/5fea696b-3ec3-47c7-bc33-212456367942-kube-api-access-k5m2m\") pod \"octavia-operator-controller-manager-69fdcfc5f5-mc6cb\" (UID: \"5fea696b-3ec3-47c7-bc33-212456367942\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.104390 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hg2s\" (UniqueName: \"kubernetes.io/projected/15fc314b-1904-488c-829c-9c5c0274ff20-kube-api-access-2hg2s\") pod \"nova-operator-controller-manager-5df598886f-dhc98\" (UID: \"15fc314b-1904-488c-829c-9c5c0274ff20\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.106332 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6lmv\" (UniqueName: \"kubernetes.io/projected/fb295ac9-43c8-4482-b557-9ac8aff33171-kube-api-access-t6lmv\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.125762 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwmxn\" (UniqueName: \"kubernetes.io/projected/e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f-kube-api-access-xwmxn\") pod \"placement-operator-controller-manager-68b6c87b68-vx59l\" (UID: \"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.125809 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjgnp\" (UniqueName: \"kubernetes.io/projected/dba339e1-27ec-4b88-aa67-b1d72aced19a-kube-api-access-fjgnp\") pod \"telemetry-operator-controller-manager-67cfc6749b-pgbzn\" (UID: \"dba339e1-27ec-4b88-aa67-b1d72aced19a\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.125913 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvdwm\" (UniqueName: \"kubernetes.io/projected/a5e07832-8614-4f4a-9d58-9fad29c4c425-kube-api-access-qvdwm\") pod \"test-operator-controller-manager-5458f77c4-nlxhj\" (UID: \"a5e07832-8614-4f4a-9d58-9fad29c4c425\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.125946 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdscf\" (UniqueName: \"kubernetes.io/projected/fdb571a6-0d71-4cd9-885d-798b4489b816-kube-api-access-vdscf\") pod \"swift-operator-controller-manager-db6d7f97b-s48sn\" (UID: \"fdb571a6-0d71-4cd9-885d-798b4489b816\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.135084 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.136349 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.143048 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lfjjf" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.144915 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.156401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwmxn\" (UniqueName: \"kubernetes.io/projected/e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f-kube-api-access-xwmxn\") pod \"placement-operator-controller-manager-68b6c87b68-vx59l\" (UID: \"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.162440 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdscf\" (UniqueName: \"kubernetes.io/projected/fdb571a6-0d71-4cd9-885d-798b4489b816-kube-api-access-vdscf\") pod \"swift-operator-controller-manager-db6d7f97b-s48sn\" (UID: \"fdb571a6-0d71-4cd9-885d-798b4489b816\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.167567 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjgnp\" (UniqueName: \"kubernetes.io/projected/dba339e1-27ec-4b88-aa67-b1d72aced19a-kube-api-access-fjgnp\") pod \"telemetry-operator-controller-manager-67cfc6749b-pgbzn\" (UID: \"dba339e1-27ec-4b88-aa67-b1d72aced19a\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.173855 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.174007 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.192534 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.205916 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.207086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.217139 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.218809 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.221860 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.222136 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-6vc7h" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.227028 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2v6l\" (UniqueName: \"kubernetes.io/projected/a7f848f3-9deb-4528-84b4-13f28c5900e9-kube-api-access-w2v6l\") pod \"watcher-operator-controller-manager-7f554bff7b-mqvhc\" (UID: \"a7f848f3-9deb-4528-84b4-13f28c5900e9\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.227088 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvdwm\" (UniqueName: \"kubernetes.io/projected/a5e07832-8614-4f4a-9d58-9fad29c4c425-kube-api-access-qvdwm\") pod \"test-operator-controller-manager-5458f77c4-nlxhj\" (UID: \"a5e07832-8614-4f4a-9d58-9fad29c4c425\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.228056 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.240072 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.248009 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.266883 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvdwm\" (UniqueName: \"kubernetes.io/projected/a5e07832-8614-4f4a-9d58-9fad29c4c425-kube-api-access-qvdwm\") pod \"test-operator-controller-manager-5458f77c4-nlxhj\" (UID: \"a5e07832-8614-4f4a-9d58-9fad29c4c425\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.280332 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.298784 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.299760 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.301469 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.304274 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-62ghw" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.311440 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.318763 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9"] Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.328085 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.328129 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2v6l\" (UniqueName: \"kubernetes.io/projected/a7f848f3-9deb-4528-84b4-13f28c5900e9-kube-api-access-w2v6l\") pod \"watcher-operator-controller-manager-7f554bff7b-mqvhc\" (UID: \"a7f848f3-9deb-4528-84b4-13f28c5900e9\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.328201 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.328239 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtsnz\" (UniqueName: \"kubernetes.io/projected/1c6a82d2-fe9c-4915-8942-865c875efb80-kube-api-access-rtsnz\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x\" (UID: \"1c6a82d2-fe9c-4915-8942-865c875efb80\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.328278 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m7qq\" (UniqueName: \"kubernetes.io/projected/087226e7-48a0-447d-9a17-59492dac054b-kube-api-access-7m7qq\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.337412 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff-cert\") pod \"infra-operator-controller-manager-656bcbd775-tfcm7\" (UID: \"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.364885 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2v6l\" (UniqueName: \"kubernetes.io/projected/a7f848f3-9deb-4528-84b4-13f28c5900e9-kube-api-access-w2v6l\") pod \"watcher-operator-controller-manager-7f554bff7b-mqvhc\" (UID: \"a7f848f3-9deb-4528-84b4-13f28c5900e9\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.409978 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.428951 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtsnz\" (UniqueName: \"kubernetes.io/projected/1c6a82d2-fe9c-4915-8942-865c875efb80-kube-api-access-rtsnz\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x\" (UID: \"1c6a82d2-fe9c-4915-8942-865c875efb80\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.429033 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m7qq\" (UniqueName: \"kubernetes.io/projected/087226e7-48a0-447d-9a17-59492dac054b-kube-api-access-7m7qq\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.429089 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.429275 5018 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.429331 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert podName:087226e7-48a0-447d-9a17-59492dac054b nodeName:}" failed. No retries permitted until 2025-10-14 07:06:08.92931297 +0000 UTC m=+985.513359597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert") pod "openstack-operator-controller-manager-7fb8c88b76-lxsvb" (UID: "087226e7-48a0-447d-9a17-59492dac054b") : secret "webhook-server-cert" not found Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.448739 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.453750 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m7qq\" (UniqueName: \"kubernetes.io/projected/087226e7-48a0-447d-9a17-59492dac054b-kube-api-access-7m7qq\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.462696 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtsnz\" (UniqueName: \"kubernetes.io/projected/1c6a82d2-fe9c-4915-8942-865c875efb80-kube-api-access-rtsnz\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x\" (UID: \"1c6a82d2-fe9c-4915-8942-865c875efb80\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.492863 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.521752 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.531398 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.536244 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fb295ac9-43c8-4482-b557-9ac8aff33171-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d448487bng8\" (UID: \"fb295ac9-43c8-4482-b557-9ac8aff33171\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.556554 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.585896 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.663997 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" Oct 14 07:06:08 crc kubenswrapper[5018]: I1014 07:06:08.944656 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.944850 5018 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 07:06:08 crc kubenswrapper[5018]: E1014 07:06:08.944944 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert podName:087226e7-48a0-447d-9a17-59492dac054b nodeName:}" failed. No retries permitted until 2025-10-14 07:06:09.944921808 +0000 UTC m=+986.528968435 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert") pod "openstack-operator-controller-manager-7fb8c88b76-lxsvb" (UID: "087226e7-48a0-447d-9a17-59492dac054b") : secret "webhook-server-cert" not found Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.029050 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.035095 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.085034 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.111988 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" event={"ID":"a81fad02-cd39-487a-ba7d-d29757b26848","Type":"ContainerStarted","Data":"33c987196091e4891c6bf39f098a366093bee68a4840cdba3a972eb89c93a717"} Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.121461 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" event={"ID":"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7","Type":"ContainerStarted","Data":"6ec5f525bc2e1def346be25b955365ec3944e9abcc7cc5d07520001337129b82"} Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.125979 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" event={"ID":"284f0e2f-1e97-4201-8313-0c604eb245c7","Type":"ContainerStarted","Data":"51d42d2875ab80e494b66fcc8871c9480ddd30a3e808b82c97055e6aa5db35b5"} Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.251828 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz"] Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.261114 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3965916_7dc3_4bfc_ae94_c57ef9b4b43c.slice/crio-63e6f12bc0353ba6ac421924d94fc19fe553a5a38bdbfb876a14e86592898700 WatchSource:0}: Error finding container 63e6f12bc0353ba6ac421924d94fc19fe553a5a38bdbfb876a14e86592898700: Status 404 returned error can't find the container with id 63e6f12bc0353ba6ac421924d94fc19fe553a5a38bdbfb876a14e86592898700 Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.262252 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.270237 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.534855 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.541563 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn"] Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.544998 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdb571a6_0d71_4cd9_885d_798b4489b816.slice/crio-c1fa47d07c20be014acef5a189c30220b3bb41c8b5207d372a3fdfc2c4d1969f WatchSource:0}: Error finding container c1fa47d07c20be014acef5a189c30220b3bb41c8b5207d372a3fdfc2c4d1969f: Status 404 returned error can't find the container with id c1fa47d07c20be014acef5a189c30220b3bb41c8b5207d372a3fdfc2c4d1969f Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.723921 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.740901 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.750581 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.769812 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.785606 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-dhc98"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.798202 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.799771 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.804026 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.815929 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l"] Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.823370 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3195f44a_f801_4f06_ae14_d83a680343a7.slice/crio-d2378fdd2457860fad23c4423c3fee72185934a4d64f020762ab692844b3e1b7 WatchSource:0}: Error finding container d2378fdd2457860fad23c4423c3fee72185934a4d64f020762ab692844b3e1b7: Status 404 returned error can't find the container with id d2378fdd2457860fad23c4423c3fee72185934a4d64f020762ab692844b3e1b7 Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.836718 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.841660 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk"] Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.841839 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5lwd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5f67fbc655-v2ppk_openstack-operators(3195f44a-f801-4f06-ae14-d83a680343a7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.845788 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode31b129a_e9ad_4f2a_a78c_04f1aeb8ac8f.slice/crio-abfad0c465cea857db30b41bd7c6d3b6e5b1a363008c6d25604c25da3edddfc4 WatchSource:0}: Error finding container abfad0c465cea857db30b41bd7c6d3b6e5b1a363008c6d25604c25da3edddfc4: Status 404 returned error can't find the container with id abfad0c465cea857db30b41bd7c6d3b6e5b1a363008c6d25604c25da3edddfc4 Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.845932 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:6889b3aa62468a9af923ac9e4f0aed418e4edd076d342c81e3a0042c3c8df19d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:bf9ab38bfe20a161e7376392f27674c833eeaaa737c2902d9132d5ee7845ab34,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:d879a4fe24c7fd9c2e7f5f91bea440c97fc7fe4891d76f761d45341e1f530b49,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:17c8a458ac0f0fe1c1f34fd5d81d0291ed16e7bfa2ced8f2a68b13dcda00702a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:121a845dacd051814fb4709fc557420363cd923a9cf2b4ed09addd394f83a3f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:8c9f2c92e8c6fcdbc2b90f8edc5e3a7055f965c93f426a70caaf2dd2808e01a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:f0e68b61dcb34ac3851ff056661b66a5f1095bc862f4391d4abcc36397a9bd49,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:37be59ccd30847c8e5a52ca48fb90e60ed1134aa96770929238f414bb9a96442,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:52c583ac7a1f249e658bb7cbc02ffe863a3e090e3baf5fc253c60e267e73c713,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:8b755bcae75f21718f07c7740080b034f4f289b859072ec2020fa0fde3f8c4f0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:bcf261baa33be891dfbda582cf679d59a943f9d311dbadbfac35f6bc3233b467,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:25f32b470fa8a112d2b3378215986adff06180f526a6ac59d892df0712ae4e92,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:0f5f6fe0a4f79c2081e0cf990940ddd8976d105c004956e09debc366cfebb7ff,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:6ed56dd94f79ff155814949be4d76a837be594c341f497191f276ffa5e681aa5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:bb5fe980813796fabb38d0ee83dc949f775aed360aea1d3a9636c9ae7342c4d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:46f7c5ad91770e99d707c09bca696e68b6248d33c0da45a89a8e2c26053b1436,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:701f27d93b40becf0d8445a0ec993eb98606d0b4fb60cf21f745d21af6ce146c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:8117021a965e52c9d2bd7b39b7ff7fa5b8217a0cd81c998a48c981191a490b6f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:1fa006cc027f654e8cc52f2ed908b1fd4de5ddd76acdaf2eff6fced95b6c3237,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:2067f20fa52f106035b36c34e1a6378ec78b90b63fac500441b8e0ef86dabf4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:2cb5b13dc65c9407d6221fd2831f310c51fb61d060cdf9b6555c9936fabad079,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:1fee90435ec29f7923062cfca6db89298f69ef7f1864df3e61258715d9d93c74,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:124d7cea22be48d4f1a8cfedec66864ccd3bea72d0fbc0d6c8e6bf4a6820e8fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:f0501133d0a13a33a8492ead797a981cf04bcd6e469de2390b645d763b9ce447,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:0f26bfcc3cc838a38a36e11055a96f7d28fb841d04aaf952494f27b1f8919d97,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:ad5f34ab2be053a1b70063f53d30c78ce1d53b7ed26d351d980256aea71d663e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:081710f3f67a74adb03d6d8f527f6ef01828243c2be24ca57436de2be8618576,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:49f308e55764cd3852a3c403107b0e5a91ccc86ec85733d4af341a18cdac4bc1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:375faec05ca0a5094bc75c17b0a775afefcc1dc30f5d7eb6ef670fb451a3fd1f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:fe04296ed96d5cae7489761537c530d175e1d306c00daf109bcd6c86bf04767f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:839f0e593dd6b59e385ec9471f4eeaa34f1c539268588114cbc34cc9a6117835,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:e46413b26dca08cdf44c265f9a0a772943a8113f67eb015185901f02bf3f467f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:fa9b8191aa63a4e2ba6b880a2680ce323afcc5ef655bff22d344edbb7bd60eff,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:76da897b5ed8dbfb9ddd60237d2bd05d9675a11e8e29054b9313f083503ab8d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:a165bb40dcffccf823656b90260c23c472338339306b1942cd0e863ac45df852,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:a5425f2dff41e2a3fc8c7168ceadc882dfb68dfd0e82e416bacd6e67604cdbac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:be240c1d8a08b9461e19d95133dda2d4aa0b37f8a22d1f54b2602a04eff217de,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:0bfbc82dc110a6453af41ead54980dd869774dd754cb9a9a4908c56b02307928,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:9483e655210e66c5c3325afee77282aecee9c2d747a6a22f1468e5b365b90a27,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:c3a430afd671404cb240b4aca54746b9c4bf93645f94c6613999196ca513913c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:8da4a885ac8adabaeb60feffda093e2670fef1b2b614bf6b04637fe221cebb9a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:94e857a625e3cc4d8ec00ba84bdcaf368cb1ecd26c5e256f998664fb063addd5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2acb369d62c3152313c3ac1d5b8c79b3e1f7d46ff9c3456e4f7e295a68f27c09,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:f7302eb8964de699cf44da13958a8ce3c1c4c05406a6fc58b6cdcb1706b8f439,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:25f6226ecbc296f4a9d3b220613400c135c9dc98dc8f06c5b26faf80cb92bb3d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:27aa5719ddf5a0130152ad03fd2a43118e42f508c83b672e3d8b01767bd66d37,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:cc67888d0b5b52501151b01a8075df22fadccc05ac31adc9c0a0135a8d99c946,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:91737b02501b73af0aee486b7447b4ae3005c904f31f1a9bf4047d0433586f80,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:6a72c6ddc7b9da9d91d37e15430acc63b6d9ebe44840d50e7fdc29eb511d3850,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:ba81e563e9ff7186a072aecff403c4e062fbaf7e06283de40fa6a724d9ff25b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:33c58faa12b90b6009f89c9c60baeadc1323b62dcb141619a7a11c3c10903560,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:85219a043b93a0cad67e3c6a21c82ebd0de298be40d2124d678e551aba9cdca3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:bcac7d43f7ca3e408eaab5af5b59bf83658e69d362b221495e3849c1412e7ac9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:b1a59a930d7b3f9411be8c55a9661adaae0927aa11d3379a93afa37e3273009f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:dc4f0d8c3ef7fc6d16df02b5e068a1a561eae9e155f4dda7664a21ded81cedca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5d3255457450c6b22246625db777c1f834ef9feacbef7d4bd1ed475a4aa65757,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:609883d4f5ee1eeeb9877b1ab3636f325970d67a6e9e0a0674a26e9c12168e5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:8257d380a013eb290d88cdbb6279e142d293115e844f8554b493a7fd64dbe5db,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:02f98ecaab5a69b32bc7184051eb1cf4ed84908766c4a78af5e8a25f3398af75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:c2ebfcd639773ddc62a37198da44de8fd76348610d91424b880d41a51b702418,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:da88acc08ce68a76a93bcc45edcd66cbef80835f4881503a9050ad67739db89b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:1f610ed4ebf657334da87dfd95b3dc5299fb3540ec1433ae3db34f0f247d8abf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:725e18e616e058867b99eb7f4b26e81374a7c6c52e1f9518392079e568d38f4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:b272e3f26af5f9f92234b6e1fd6ce860f46308e5cbe38cdbf631d19499ae115f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:fcff3afee08fff84ac33b5c25e3ee22c9b05402624163b1b2b9bb2e39ed7205d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:f42888dd0e4efdc82c334650e97164c6b50e195acb4eee0bc77b35337ccc1328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:97feaea1e66145857f3eb548d741ee56062b97fd3e8f4d136a5ca807c49c0cca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:0eb4330fbe649411e9de5bc7a80af6050d765283f97d46a725ee49269b0600cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:b222aa9f56abead1e1b5d8159c0bc3687024d151f53a33cafdf33377feab8f35,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:74280842b5e733f8456ae433560fc90aca37fd6ccc0c6d902b835d36b978e041,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:9b2e302646bbf02dd583d9c340c5e7cdf708241c4c32700ebe7696ee235725bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:74db5648836429519c8bebf9d1e7ea4dd3961c7b3b83bb3d50dffbcfe4c41a5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:062ac305d19a6eec23a678835b5aa3dc3dda82c7470089408f98f5728e90ce5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:c83af60e124c22c9e9f2cdf6f1f00bd63a0d9b038dafac29c1c91c5c978923bc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:00176d96a79fb4eda8540cb22f1dce081234c60e9e6a2b034950bc8d6e718b79,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6lmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-55b7d448487bng8_openstack-operators(fb295ac9-43c8-4482-b557-9ac8aff33171): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.846446 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fea696b_3ec3_47c7_bc33_212456367942.slice/crio-33a28a7c38cc6e5dde698fc9932088f00ddc66fa8af9c7061c0642b7ef855bcf WatchSource:0}: Error finding container 33a28a7c38cc6e5dde698fc9932088f00ddc66fa8af9c7061c0642b7ef855bcf: Status 404 returned error can't find the container with id 33a28a7c38cc6e5dde698fc9932088f00ddc66fa8af9c7061c0642b7ef855bcf Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.850417 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x"] Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.854361 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7"] Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.855239 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf3020fc_9cba_409f_8355_c76ddfb47db2.slice/crio-4ef8298f9debe081cfc477f900dc0dfe690e1dd27ff172fbe8b8d2fab1c39e6a WatchSource:0}: Error finding container 4ef8298f9debe081cfc477f900dc0dfe690e1dd27ff172fbe8b8d2fab1c39e6a: Status 404 returned error can't find the container with id 4ef8298f9debe081cfc477f900dc0dfe690e1dd27ff172fbe8b8d2fab1c39e6a Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.855400 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c6a82d2_fe9c_4915_8942_865c875efb80.slice/crio-c6ba7ce5dba9c77d7f681d53eaf87a9b206418f52bee0f2471ea49d4ee767e13 WatchSource:0}: Error finding container c6ba7ce5dba9c77d7f681d53eaf87a9b206418f52bee0f2471ea49d4ee767e13: Status 404 returned error can't find the container with id c6ba7ce5dba9c77d7f681d53eaf87a9b206418f52bee0f2471ea49d4ee767e13 Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.856069 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xwmxn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-vx59l_openstack-operators(e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.857086 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5m2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69fdcfc5f5-mc6cb_openstack-operators(5fea696b-3ec3-47c7-bc33-212456367942): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.857090 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rtsnz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x_openstack-operators(1c6a82d2-fe9c-4915-8942-865c875efb80): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.857197 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n44m8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-79df5fb58c-wltkk_openstack-operators(df3020fc-9cba-409f-8355-c76ddfb47db2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.859155 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qlg87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-tfcm7_openstack-operators(dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:06:09 crc kubenswrapper[5018]: E1014 07:06:09.859252 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" podUID="1c6a82d2-fe9c-4915-8942-865c875efb80" Oct 14 07:06:09 crc kubenswrapper[5018]: W1014 07:06:09.883236 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7f848f3_9deb_4528_84b4_13f28c5900e9.slice/crio-f130b3c923833955a7d732c560b9bab3299931700f429b32284e22bc770e9879 WatchSource:0}: Error finding container f130b3c923833955a7d732c560b9bab3299931700f429b32284e22bc770e9879: Status 404 returned error can't find the container with id f130b3c923833955a7d732c560b9bab3299931700f429b32284e22bc770e9879 Oct 14 07:06:09 crc kubenswrapper[5018]: I1014 07:06:09.976283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.005151 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/087226e7-48a0-447d-9a17-59492dac054b-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-lxsvb\" (UID: \"087226e7-48a0-447d-9a17-59492dac054b\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.081205 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" podUID="fb295ac9-43c8-4482-b557-9ac8aff33171" Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.087580 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" podUID="3195f44a-f801-4f06-ae14-d83a680343a7" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.120160 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.153012 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" event={"ID":"a5e07832-8614-4f4a-9d58-9fad29c4c425","Type":"ContainerStarted","Data":"151d73fec2df2f82cc94608ecaf1c40a9eb58e3d9b49519357ea1c742be49495"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.162646 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" event={"ID":"88f2a077-948e-4245-834d-9f943670b40f","Type":"ContainerStarted","Data":"168cf6166083dded499b75e001831576e148544ddac108142e9d3819de6e13a9"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.168249 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" event={"ID":"df3020fc-9cba-409f-8355-c76ddfb47db2","Type":"ContainerStarted","Data":"4ef8298f9debe081cfc477f900dc0dfe690e1dd27ff172fbe8b8d2fab1c39e6a"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.189613 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" event={"ID":"0e01bea2-81b7-4b83-881c-c40f3a8bd306","Type":"ContainerStarted","Data":"946d0c2ea271d47b865ef1c55db9bf791a31833d10b1f209a5259c17e5799a70"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.191042 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" event={"ID":"fb295ac9-43c8-4482-b557-9ac8aff33171","Type":"ContainerStarted","Data":"00c582c71802215f2507ef8f776c8f50336192a011b6978db1c0004d11a54bed"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.191071 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" event={"ID":"fb295ac9-43c8-4482-b557-9ac8aff33171","Type":"ContainerStarted","Data":"48c20022b2dc8e46a8b32cc101a8f6ac16070f0d389210425209254637fc6156"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.193011 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" podUID="fb295ac9-43c8-4482-b557-9ac8aff33171" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.198631 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" event={"ID":"4afa0a87-9ea4-4d10-857d-8c046132dbd3","Type":"ContainerStarted","Data":"ed8ce79de9d676c6d0ce44e6446f275d836cf0bfb61815ece1a2006449aa4f34"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.211168 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" event={"ID":"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f","Type":"ContainerStarted","Data":"abfad0c465cea857db30b41bd7c6d3b6e5b1a363008c6d25604c25da3edddfc4"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.224049 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" event={"ID":"6bf6a59f-81bd-4003-a387-32c703648d21","Type":"ContainerStarted","Data":"0ffbfad11549167ac6c1b62df8d425f4cdee530c223f60745d43b8850e66a34a"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.233073 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" event={"ID":"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c","Type":"ContainerStarted","Data":"63e6f12bc0353ba6ac421924d94fc19fe553a5a38bdbfb876a14e86592898700"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.234007 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" event={"ID":"5fea696b-3ec3-47c7-bc33-212456367942","Type":"ContainerStarted","Data":"33a28a7c38cc6e5dde698fc9932088f00ddc66fa8af9c7061c0642b7ef855bcf"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.265236 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" podUID="5fea696b-3ec3-47c7-bc33-212456367942" Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.272995 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" podUID="df3020fc-9cba-409f-8355-c76ddfb47db2" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.273210 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" event={"ID":"3195f44a-f801-4f06-ae14-d83a680343a7","Type":"ContainerStarted","Data":"ffab7e68b8e1fbe94417644cb110b5a3d48599d8e377524fb25bb7e6c600a899"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.273235 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" event={"ID":"3195f44a-f801-4f06-ae14-d83a680343a7","Type":"ContainerStarted","Data":"d2378fdd2457860fad23c4423c3fee72185934a4d64f020762ab692844b3e1b7"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.276993 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" podUID="3195f44a-f801-4f06-ae14-d83a680343a7" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.298675 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" event={"ID":"a7f848f3-9deb-4528-84b4-13f28c5900e9","Type":"ContainerStarted","Data":"f130b3c923833955a7d732c560b9bab3299931700f429b32284e22bc770e9879"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.315077 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" podUID="e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.319839 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" event={"ID":"d94d1327-34a3-45fe-963c-b80cc8931811","Type":"ContainerStarted","Data":"485f3dfbd54a25438bc6c6139cc7c77ad6814b1b709ab94c19574995ae4ce4b7"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.339918 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" event={"ID":"1c6a82d2-fe9c-4915-8942-865c875efb80","Type":"ContainerStarted","Data":"c6ba7ce5dba9c77d7f681d53eaf87a9b206418f52bee0f2471ea49d4ee767e13"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.348656 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" podUID="dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff" Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.348826 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" podUID="1c6a82d2-fe9c-4915-8942-865c875efb80" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.356888 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" event={"ID":"fdb571a6-0d71-4cd9-885d-798b4489b816","Type":"ContainerStarted","Data":"c1fa47d07c20be014acef5a189c30220b3bb41c8b5207d372a3fdfc2c4d1969f"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.369020 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" event={"ID":"15fc314b-1904-488c-829c-9c5c0274ff20","Type":"ContainerStarted","Data":"6b78554b598c1108f5db625fd17b4922f15bed35e469c70172fd074f55843f75"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.371027 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" event={"ID":"3c93e5a0-bd99-4348-8562-6584ca640fc8","Type":"ContainerStarted","Data":"b8e4f06721637d5412a2649117d8b8c2839955da38a22645ec497215dd7f44b8"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.377140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" event={"ID":"dba339e1-27ec-4b88-aa67-b1d72aced19a","Type":"ContainerStarted","Data":"fde0402f04611186a6fd99409dbf9a9e51e5721dd0967cf0e9bbd73fe93cd5f5"} Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.395940 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" event={"ID":"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff","Type":"ContainerStarted","Data":"66783f17ba8bcccedd143e4b829f7f62a7b88ae14e420ace2be380b86f81aeec"} Oct 14 07:06:10 crc kubenswrapper[5018]: E1014 07:06:10.406717 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" podUID="dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff" Oct 14 07:06:10 crc kubenswrapper[5018]: I1014 07:06:10.839606 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb"] Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.409962 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" event={"ID":"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f","Type":"ContainerStarted","Data":"8e01e04a1449dbef9925edf93cd28fb3c0d75f1761a90fb2437eec9df471ffc1"} Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.413392 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" podUID="e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f" Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.420082 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" event={"ID":"5fea696b-3ec3-47c7-bc33-212456367942","Type":"ContainerStarted","Data":"8281dce091d100a1c290dfe8936b1363d31c29640ec6987e6cc3b486979a0064"} Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.430242 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" podUID="5fea696b-3ec3-47c7-bc33-212456367942" Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.450041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" event={"ID":"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff","Type":"ContainerStarted","Data":"6b61f3ac1bada0d4cf7f669f1808347a194ea3ac870ac86a2891112539b94ea8"} Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.456772 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" podUID="dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff" Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.462168 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" event={"ID":"087226e7-48a0-447d-9a17-59492dac054b","Type":"ContainerStarted","Data":"8ae890d0e4f168cd21759bd642a0e58d3f01623868828e0a46bb67b0825fa930"} Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.462209 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" event={"ID":"087226e7-48a0-447d-9a17-59492dac054b","Type":"ContainerStarted","Data":"90b1d72fd7ce4634eae769f1b49a56d22ad662830f7b19884692fb25c40c6747"} Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.462218 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" event={"ID":"087226e7-48a0-447d-9a17-59492dac054b","Type":"ContainerStarted","Data":"05a36d19ea0a1f0387d545939b4b9a0eb50f1e2213fa0428a59a86ca946765db"} Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.462808 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.482684 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" event={"ID":"df3020fc-9cba-409f-8355-c76ddfb47db2","Type":"ContainerStarted","Data":"7130e32a73e38112498660d251779e6d0076a66fc4f2257c5f9899fe0dbae27a"} Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.486828 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" podUID="df3020fc-9cba-409f-8355-c76ddfb47db2" Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.486902 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" podUID="1c6a82d2-fe9c-4915-8942-865c875efb80" Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.486918 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" podUID="3195f44a-f801-4f06-ae14-d83a680343a7" Oct 14 07:06:11 crc kubenswrapper[5018]: E1014 07:06:11.486995 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" podUID="fb295ac9-43c8-4482-b557-9ac8aff33171" Oct 14 07:06:11 crc kubenswrapper[5018]: I1014 07:06:11.518797 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" podStartSLOduration=3.518785712 podStartE2EDuration="3.518785712s" podCreationTimestamp="2025-10-14 07:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:06:11.516133448 +0000 UTC m=+988.100180075" watchObservedRunningTime="2025-10-14 07:06:11.518785712 +0000 UTC m=+988.102832339" Oct 14 07:06:12 crc kubenswrapper[5018]: E1014 07:06:12.499233 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" podUID="e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f" Oct 14 07:06:12 crc kubenswrapper[5018]: E1014 07:06:12.499264 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" podUID="df3020fc-9cba-409f-8355-c76ddfb47db2" Oct 14 07:06:12 crc kubenswrapper[5018]: E1014 07:06:12.499265 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" podUID="dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff" Oct 14 07:06:12 crc kubenswrapper[5018]: E1014 07:06:12.499348 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" podUID="5fea696b-3ec3-47c7-bc33-212456367942" Oct 14 07:06:20 crc kubenswrapper[5018]: I1014 07:06:20.127152 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-lxsvb" Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.591197 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" event={"ID":"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c","Type":"ContainerStarted","Data":"2420407dd4cf160d629012397b7593af9ababbc20059951678c4cd1241b58d3f"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.591654 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" event={"ID":"a3965916-7dc3-4bfc-ae94-c57ef9b4b43c","Type":"ContainerStarted","Data":"48caf7116031a897906afe57ea6f0a503e41622f31673a690f6e5e3820e8d63c"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.592291 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.597210 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" event={"ID":"3c93e5a0-bd99-4348-8562-6584ca640fc8","Type":"ContainerStarted","Data":"114cd1f08ddc646b3d20ee92ae8afaf5c39c19a030e16a0bae1f9aa33b3eb880"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.603596 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" event={"ID":"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7","Type":"ContainerStarted","Data":"9753a62bc499c7ef93f52cea7238bfc3040afde9fdcb7aeff5d01833aec806a2"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.604693 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" event={"ID":"a7f848f3-9deb-4528-84b4-13f28c5900e9","Type":"ContainerStarted","Data":"4e8f63ee2c97273968f0da6b2c01bb6432f81922247caee5fabed0a3bf095f6f"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.605890 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" event={"ID":"fdb571a6-0d71-4cd9-885d-798b4489b816","Type":"ContainerStarted","Data":"9b5d23e427b3579f41a59ec6784233819d9ff86cd15a742d8885c017ec8afcb1"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.610591 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" event={"ID":"6bf6a59f-81bd-4003-a387-32c703648d21","Type":"ContainerStarted","Data":"6d8720df3a6285675b43a5ce60c9ab3a3df60b7c43212ad6e93ccdfc230794c1"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.623088 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" event={"ID":"88f2a077-948e-4245-834d-9f943670b40f","Type":"ContainerStarted","Data":"94d118a83fbb8435d1f9752d4b804cb709840e5b3866e89ffbed678a23344787"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.631105 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" event={"ID":"d94d1327-34a3-45fe-963c-b80cc8931811","Type":"ContainerStarted","Data":"9e9135de4c2c4755c53426153684dbf7d582a61ed17a46c0590004ea7b9a9fbf"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.632402 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" event={"ID":"4afa0a87-9ea4-4d10-857d-8c046132dbd3","Type":"ContainerStarted","Data":"dd00fc2bfd7dba9c69facd06ab06a9f4c8a4e4f2f6f025802f5aaa2ef72a37b0"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.633525 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" event={"ID":"a5e07832-8614-4f4a-9d58-9fad29c4c425","Type":"ContainerStarted","Data":"afeb2f332da48e0dd323abe00d7d8fee7f79b66372e7955944543845a4a2b7d9"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.652376 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" event={"ID":"284f0e2f-1e97-4201-8313-0c604eb245c7","Type":"ContainerStarted","Data":"98d849ae338bd0b8d367de1833ea1036e7b622073d38bcc74da4addf485dcf15"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.662355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" event={"ID":"dba339e1-27ec-4b88-aa67-b1d72aced19a","Type":"ContainerStarted","Data":"85a011d63e108486c8b127a66cc9d60b156e4952b5021f1528dba911996dd488"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.664948 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" event={"ID":"0e01bea2-81b7-4b83-881c-c40f3a8bd306","Type":"ContainerStarted","Data":"c8bea371d1d93ef3b3d638b98aec87c048956175c213186c42db6da1bc4a935b"} Oct 14 07:06:21 crc kubenswrapper[5018]: I1014 07:06:21.668449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" event={"ID":"15fc314b-1904-488c-829c-9c5c0274ff20","Type":"ContainerStarted","Data":"1e3a71a325236f96e9069ae900caf4c0b99655878c78e1543f78986b8b3edd38"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.677826 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" event={"ID":"dba339e1-27ec-4b88-aa67-b1d72aced19a","Type":"ContainerStarted","Data":"610c7de31cc7867c423cc39dae8b361841b7b481a445b4a507bf35f7803ba28c"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.678762 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.681254 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" event={"ID":"a5e07832-8614-4f4a-9d58-9fad29c4c425","Type":"ContainerStarted","Data":"c21cfbd9ce393e8eed1f2b09390c24b8fbb137d94ca27d9191ce45ef4e99b196"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.681410 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.683357 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" event={"ID":"a0e4cf91-8413-4a0f-b5ec-06f8becca6a7","Type":"ContainerStarted","Data":"1c88b163f201357aa13796d808fd8c28155ebf8e59f80c0dc58336a5f1164d29"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.683473 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.684613 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" event={"ID":"a7f848f3-9deb-4528-84b4-13f28c5900e9","Type":"ContainerStarted","Data":"7a9dcb64430ab839971b93b5d0180f9d25a7d2d72a56e5a24c315fb048226222"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.684815 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.686758 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" event={"ID":"284f0e2f-1e97-4201-8313-0c604eb245c7","Type":"ContainerStarted","Data":"fc333359f1074f0163410056fda9c443a356cc7f5ecdcf38737b0d5e009965b5"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.686898 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.688358 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" event={"ID":"4afa0a87-9ea4-4d10-857d-8c046132dbd3","Type":"ContainerStarted","Data":"beda653aa14b1abc6f15377efb8e86b1cba454258f2b9963750eedd7d3c7b8d1"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.688461 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.690881 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" event={"ID":"15fc314b-1904-488c-829c-9c5c0274ff20","Type":"ContainerStarted","Data":"93a6cd6102c27897cdffde5018a66326dc75f8b30becc0be4dda517718c7c263"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.691239 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.692647 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" event={"ID":"3c93e5a0-bd99-4348-8562-6584ca640fc8","Type":"ContainerStarted","Data":"c12918a1ed46ec571d12f7543ce440c51d011cb0cbb7150b7983593ae9752eb2"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.692824 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.694793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" event={"ID":"a81fad02-cd39-487a-ba7d-d29757b26848","Type":"ContainerStarted","Data":"b42ced59de96f53ef643e086819e653dbee9fc34de0c4586a73891fb21617487"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.694816 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" event={"ID":"a81fad02-cd39-487a-ba7d-d29757b26848","Type":"ContainerStarted","Data":"f4298c3205ccd0d49d24cebe23c312590a7d11b37addc18e1c09853557d7974b"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.696885 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" event={"ID":"88f2a077-948e-4245-834d-9f943670b40f","Type":"ContainerStarted","Data":"029cc4a5ac5344267fcd8e1c26df6df0abc0c8860f1eabbeb7338b60c064dcb4"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.697011 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.701631 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" podStartSLOduration=4.296344109 podStartE2EDuration="15.701600182s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.265797816 +0000 UTC m=+985.849844443" lastFinishedPulling="2025-10-14 07:06:20.671053889 +0000 UTC m=+997.255100516" observedRunningTime="2025-10-14 07:06:21.646039668 +0000 UTC m=+998.230086315" watchObservedRunningTime="2025-10-14 07:06:22.701600182 +0000 UTC m=+999.285646819" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.702329 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" podStartSLOduration=4.517158417 podStartE2EDuration="15.702323583s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.547971305 +0000 UTC m=+986.132017932" lastFinishedPulling="2025-10-14 07:06:20.733136471 +0000 UTC m=+997.317183098" observedRunningTime="2025-10-14 07:06:22.697398915 +0000 UTC m=+999.281445552" watchObservedRunningTime="2025-10-14 07:06:22.702323583 +0000 UTC m=+999.286370220" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.702844 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" event={"ID":"6bf6a59f-81bd-4003-a387-32c703648d21","Type":"ContainerStarted","Data":"4a987cb0517b999820d3599a19eedcf148e60579ca1359e8e0d76c4008a3b998"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.702983 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.711498 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" event={"ID":"d94d1327-34a3-45fe-963c-b80cc8931811","Type":"ContainerStarted","Data":"4d96a818c18ff77682bc901240524f1e96ee33f6e52d52ffced4bbd90127dff4"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.711770 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.713481 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" event={"ID":"0e01bea2-81b7-4b83-881c-c40f3a8bd306","Type":"ContainerStarted","Data":"391deb53ba1a0c69af67a9ba113003e7a0ebedf1316d3832ac20ded15836d336"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.714531 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.716700 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" event={"ID":"fdb571a6-0d71-4cd9-885d-798b4489b816","Type":"ContainerStarted","Data":"286cd27a82340c0da6f9a6c454591d9083071a5a66713f80d66e680d3caae8dd"} Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.727822 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" podStartSLOduration=4.866233911 podStartE2EDuration="15.727797423s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.821035019 +0000 UTC m=+986.405081646" lastFinishedPulling="2025-10-14 07:06:20.682598521 +0000 UTC m=+997.266645158" observedRunningTime="2025-10-14 07:06:22.719262655 +0000 UTC m=+999.303309282" watchObservedRunningTime="2025-10-14 07:06:22.727797423 +0000 UTC m=+999.311844070" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.735903 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" podStartSLOduration=4.172844735 podStartE2EDuration="15.735891039s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.077567367 +0000 UTC m=+985.661613994" lastFinishedPulling="2025-10-14 07:06:20.640613631 +0000 UTC m=+997.224660298" observedRunningTime="2025-10-14 07:06:22.731735433 +0000 UTC m=+999.315782060" watchObservedRunningTime="2025-10-14 07:06:22.735891039 +0000 UTC m=+999.319937666" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.752906 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" podStartSLOduration=4.3805205560000005 podStartE2EDuration="15.752890393s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.269106198 +0000 UTC m=+985.853152825" lastFinishedPulling="2025-10-14 07:06:20.641476015 +0000 UTC m=+997.225522662" observedRunningTime="2025-10-14 07:06:22.745459455 +0000 UTC m=+999.329506082" watchObservedRunningTime="2025-10-14 07:06:22.752890393 +0000 UTC m=+999.336937030" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.776552 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" podStartSLOduration=4.021927722 podStartE2EDuration="14.776539752s" podCreationTimestamp="2025-10-14 07:06:08 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.886072263 +0000 UTC m=+986.470118890" lastFinishedPulling="2025-10-14 07:06:20.640684283 +0000 UTC m=+997.224730920" observedRunningTime="2025-10-14 07:06:22.774284189 +0000 UTC m=+999.358330816" watchObservedRunningTime="2025-10-14 07:06:22.776539752 +0000 UTC m=+999.360586379" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.778111 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" podStartSLOduration=4.266427985 podStartE2EDuration="15.778106376s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.087176295 +0000 UTC m=+985.671222922" lastFinishedPulling="2025-10-14 07:06:20.598854666 +0000 UTC m=+997.182901313" observedRunningTime="2025-10-14 07:06:22.757945874 +0000 UTC m=+999.341992491" watchObservedRunningTime="2025-10-14 07:06:22.778106376 +0000 UTC m=+999.362153003" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.797721 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" podStartSLOduration=4.880383675 podStartE2EDuration="15.797705262s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.752948811 +0000 UTC m=+986.336995438" lastFinishedPulling="2025-10-14 07:06:20.670270348 +0000 UTC m=+997.254317025" observedRunningTime="2025-10-14 07:06:22.796456348 +0000 UTC m=+999.380502995" watchObservedRunningTime="2025-10-14 07:06:22.797705262 +0000 UTC m=+999.381751889" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.817096 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" podStartSLOduration=4.976785524 podStartE2EDuration="15.817077543s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.779585623 +0000 UTC m=+986.363632250" lastFinishedPulling="2025-10-14 07:06:20.619877632 +0000 UTC m=+997.203924269" observedRunningTime="2025-10-14 07:06:22.812220187 +0000 UTC m=+999.396266814" watchObservedRunningTime="2025-10-14 07:06:22.817077543 +0000 UTC m=+999.401124170" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.838393 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" podStartSLOduration=4.942146838 podStartE2EDuration="15.838367206s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.745307448 +0000 UTC m=+986.329354065" lastFinishedPulling="2025-10-14 07:06:20.641527766 +0000 UTC m=+997.225574433" observedRunningTime="2025-10-14 07:06:22.835882357 +0000 UTC m=+999.419928984" watchObservedRunningTime="2025-10-14 07:06:22.838367206 +0000 UTC m=+999.422413863" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.864668 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" podStartSLOduration=3.52609186 podStartE2EDuration="15.864638699s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:08.393773069 +0000 UTC m=+984.977819686" lastFinishedPulling="2025-10-14 07:06:20.732319898 +0000 UTC m=+997.316366525" observedRunningTime="2025-10-14 07:06:22.853858748 +0000 UTC m=+999.437905385" watchObservedRunningTime="2025-10-14 07:06:22.864638699 +0000 UTC m=+999.448685336" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.893578 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" podStartSLOduration=4.47207944 podStartE2EDuration="15.893563496s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.209203278 +0000 UTC m=+985.793249905" lastFinishedPulling="2025-10-14 07:06:20.630687304 +0000 UTC m=+997.214733961" observedRunningTime="2025-10-14 07:06:22.888260868 +0000 UTC m=+999.472307505" watchObservedRunningTime="2025-10-14 07:06:22.893563496 +0000 UTC m=+999.477610123" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.903851 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" podStartSLOduration=4.810195288 podStartE2EDuration="15.903840092s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.546959706 +0000 UTC m=+986.131006333" lastFinishedPulling="2025-10-14 07:06:20.6406045 +0000 UTC m=+997.224651137" observedRunningTime="2025-10-14 07:06:22.902937687 +0000 UTC m=+999.486984334" watchObservedRunningTime="2025-10-14 07:06:22.903840092 +0000 UTC m=+999.487886719" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.923084 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" podStartSLOduration=5.065837887 podStartE2EDuration="15.923064888s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.813634203 +0000 UTC m=+986.397680830" lastFinishedPulling="2025-10-14 07:06:20.670861194 +0000 UTC m=+997.254907831" observedRunningTime="2025-10-14 07:06:22.921759672 +0000 UTC m=+999.505806309" watchObservedRunningTime="2025-10-14 07:06:22.923064888 +0000 UTC m=+999.507111525" Oct 14 07:06:22 crc kubenswrapper[5018]: I1014 07:06:22.941987 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" podStartSLOduration=4.554432615 podStartE2EDuration="15.941970645s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.275915148 +0000 UTC m=+985.859961765" lastFinishedPulling="2025-10-14 07:06:20.663453178 +0000 UTC m=+997.247499795" observedRunningTime="2025-10-14 07:06:22.940323409 +0000 UTC m=+999.524370046" watchObservedRunningTime="2025-10-14 07:06:22.941970645 +0000 UTC m=+999.526017282" Oct 14 07:06:23 crc kubenswrapper[5018]: I1014 07:06:23.730461 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:23 crc kubenswrapper[5018]: I1014 07:06:23.732258 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:25 crc kubenswrapper[5018]: I1014 07:06:25.742480 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" event={"ID":"e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f","Type":"ContainerStarted","Data":"dc15808c250c419dae6957f9ff5224ec5d365fd6da9a9151e79654c0d12f92f5"} Oct 14 07:06:25 crc kubenswrapper[5018]: I1014 07:06:25.743473 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:25 crc kubenswrapper[5018]: I1014 07:06:25.763508 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" podStartSLOduration=3.369177444 podStartE2EDuration="18.763485685s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.855972584 +0000 UTC m=+986.440019211" lastFinishedPulling="2025-10-14 07:06:25.250280825 +0000 UTC m=+1001.834327452" observedRunningTime="2025-10-14 07:06:25.75899789 +0000 UTC m=+1002.343044527" watchObservedRunningTime="2025-10-14 07:06:25.763485685 +0000 UTC m=+1002.347532312" Oct 14 07:06:26 crc kubenswrapper[5018]: I1014 07:06:26.751428 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" event={"ID":"fb295ac9-43c8-4482-b557-9ac8aff33171","Type":"ContainerStarted","Data":"f6a1ec653922fa5d6b956976e851bd267a1012bad8326dd1ab698ae7c84e5b14"} Oct 14 07:06:26 crc kubenswrapper[5018]: I1014 07:06:26.752562 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.791840 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-swhnp" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.822047 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" podStartSLOduration=4.264906394 podStartE2EDuration="20.82202561s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.845507152 +0000 UTC m=+986.429553779" lastFinishedPulling="2025-10-14 07:06:26.402626368 +0000 UTC m=+1002.986672995" observedRunningTime="2025-10-14 07:06:26.779073096 +0000 UTC m=+1003.363119723" watchObservedRunningTime="2025-10-14 07:06:27.82202561 +0000 UTC m=+1004.406072237" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.822834 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-vwpjg" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.847741 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-4fgw9" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.873973 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-c6q56" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.891757 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gnt2q" Oct 14 07:06:27 crc kubenswrapper[5018]: I1014 07:06:27.921478 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-b9n5t" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.177300 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-pgbzn" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.178995 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-jrnfz" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.195939 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-wrxs4" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.247701 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-92pt5" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.253690 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-rfcrs" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.254104 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-dhc98" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.453274 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-s48sn" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.495634 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5458f77c4-nlxhj" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.531262 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-mqvhc" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.769732 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" event={"ID":"dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff","Type":"ContainerStarted","Data":"363d90f5b596e85c15f79254dd2b477255b1a3824fc31ffdb964e3fe2b55e67a"} Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.769911 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.772389 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" event={"ID":"1c6a82d2-fe9c-4915-8942-865c875efb80","Type":"ContainerStarted","Data":"c36fb64f95e75b98e9540550e1fd47c87bc7aaeb7399429f1395e4bae5d326f6"} Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.794286 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" podStartSLOduration=3.768633044 podStartE2EDuration="21.794270682s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.858897935 +0000 UTC m=+986.442944562" lastFinishedPulling="2025-10-14 07:06:27.884535573 +0000 UTC m=+1004.468582200" observedRunningTime="2025-10-14 07:06:28.793016207 +0000 UTC m=+1005.377062844" watchObservedRunningTime="2025-10-14 07:06:28.794270682 +0000 UTC m=+1005.378317319" Oct 14 07:06:28 crc kubenswrapper[5018]: I1014 07:06:28.813022 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x" podStartSLOduration=2.798270651 podStartE2EDuration="20.813005384s" podCreationTimestamp="2025-10-14 07:06:08 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.856979372 +0000 UTC m=+986.441025999" lastFinishedPulling="2025-10-14 07:06:27.871714105 +0000 UTC m=+1004.455760732" observedRunningTime="2025-10-14 07:06:28.808687124 +0000 UTC m=+1005.392733761" watchObservedRunningTime="2025-10-14 07:06:28.813005384 +0000 UTC m=+1005.397052021" Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.783842 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" event={"ID":"5fea696b-3ec3-47c7-bc33-212456367942","Type":"ContainerStarted","Data":"2ca384608ce8f7612087ca95000633aa2165d67e5694596436a2f7878707b637"} Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.785547 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.795282 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" event={"ID":"3195f44a-f801-4f06-ae14-d83a680343a7","Type":"ContainerStarted","Data":"135cea8a1aa7290c921d56aec33caa1b1353dab50e25d6beaba90ed426e9fc12"} Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.796124 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.809993 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" podStartSLOduration=3.296592811 podStartE2EDuration="22.809973925s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.856966081 +0000 UTC m=+986.441012708" lastFinishedPulling="2025-10-14 07:06:29.370347195 +0000 UTC m=+1005.954393822" observedRunningTime="2025-10-14 07:06:29.80621826 +0000 UTC m=+1006.390264927" watchObservedRunningTime="2025-10-14 07:06:29.809973925 +0000 UTC m=+1006.394020562" Oct 14 07:06:29 crc kubenswrapper[5018]: I1014 07:06:29.825951 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" podStartSLOduration=3.298840533 podStartE2EDuration="22.82593397s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.841693895 +0000 UTC m=+986.425740522" lastFinishedPulling="2025-10-14 07:06:29.368787332 +0000 UTC m=+1005.952833959" observedRunningTime="2025-10-14 07:06:29.822110973 +0000 UTC m=+1006.406157610" watchObservedRunningTime="2025-10-14 07:06:29.82593397 +0000 UTC m=+1006.409980607" Oct 14 07:06:30 crc kubenswrapper[5018]: I1014 07:06:30.805573 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" event={"ID":"df3020fc-9cba-409f-8355-c76ddfb47db2","Type":"ContainerStarted","Data":"99335e9515df6eb737b1c21cacc0b00ed28ade6cffdbb97fb035379645133d01"} Oct 14 07:06:30 crc kubenswrapper[5018]: I1014 07:06:30.806283 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:30 crc kubenswrapper[5018]: I1014 07:06:30.876046 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" podStartSLOduration=3.624209136 podStartE2EDuration="23.876022062s" podCreationTimestamp="2025-10-14 07:06:07 +0000 UTC" firstStartedPulling="2025-10-14 07:06:09.857136996 +0000 UTC m=+986.441183623" lastFinishedPulling="2025-10-14 07:06:30.108949922 +0000 UTC m=+1006.692996549" observedRunningTime="2025-10-14 07:06:30.871814435 +0000 UTC m=+1007.455861122" watchObservedRunningTime="2025-10-14 07:06:30.876022062 +0000 UTC m=+1007.460068729" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.211306 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-v2ppk" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.291087 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-mc6cb" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.306741 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-wltkk" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.414548 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-vx59l" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.566781 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d448487bng8" Oct 14 07:06:38 crc kubenswrapper[5018]: I1014 07:06:38.594724 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tfcm7" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.884441 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.890585 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.900098 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.900603 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.900638 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.900928 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-xskkn" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.907140 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.954350 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.956125 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.959208 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 07:06:55 crc kubenswrapper[5018]: I1014 07:06:55.965956 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.047552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.047828 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb79m\" (UniqueName: \"kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.149665 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.149753 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.149790 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb79m\" (UniqueName: \"kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.149810 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fql6b\" (UniqueName: \"kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.149890 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.150683 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.170457 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb79m\" (UniqueName: \"kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m\") pod \"dnsmasq-dns-5d487d97d7-9xpdm\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.226260 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.251878 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.252246 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.252410 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fql6b\" (UniqueName: \"kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.252555 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.254030 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.286356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fql6b\" (UniqueName: \"kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b\") pod \"dnsmasq-dns-6948694bd9-8hs6c\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.578181 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:06:56 crc kubenswrapper[5018]: I1014 07:06:56.688929 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:06:57 crc kubenswrapper[5018]: I1014 07:06:57.066838 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" event={"ID":"331cc27c-304b-4264-a43f-4a03657c4435","Type":"ContainerStarted","Data":"2f31c88923ee234e057545479a668dfc61fec238889f712707a86c15339a3712"} Oct 14 07:06:57 crc kubenswrapper[5018]: I1014 07:06:57.080821 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:06:57 crc kubenswrapper[5018]: W1014 07:06:57.088800 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3da82b5_168a_4504_b171_6d44e63f4d3b.slice/crio-351974fd17b1d860a44337f00ac7d52e365d93a88aaac717d5a29e2c0f12f1b1 WatchSource:0}: Error finding container 351974fd17b1d860a44337f00ac7d52e365d93a88aaac717d5a29e2c0f12f1b1: Status 404 returned error can't find the container with id 351974fd17b1d860a44337f00ac7d52e365d93a88aaac717d5a29e2c0f12f1b1 Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.079136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" event={"ID":"e3da82b5-168a-4504-b171-6d44e63f4d3b","Type":"ContainerStarted","Data":"351974fd17b1d860a44337f00ac7d52e365d93a88aaac717d5a29e2c0f12f1b1"} Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.561949 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.587548 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.589112 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.600789 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.786417 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.786770 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.786817 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6kwp\" (UniqueName: \"kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.830783 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.858160 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.859551 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.880363 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.891213 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.891270 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8krps\" (UniqueName: \"kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.891984 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.892040 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.892091 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6kwp\" (UniqueName: \"kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.892120 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.892991 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.893441 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.917838 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6kwp\" (UniqueName: \"kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp\") pod \"dnsmasq-dns-86f694bf-l2hnw\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.922437 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.994405 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.994461 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.994489 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8krps\" (UniqueName: \"kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.995748 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:58 crc kubenswrapper[5018]: I1014 07:06:58.995955 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.022787 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8krps\" (UniqueName: \"kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps\") pod \"dnsmasq-dns-7869c47d6c-fwmld\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.191119 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.210705 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:06:59 crc kubenswrapper[5018]: W1014 07:06:59.218139 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3255895_3e18_419b_a8fc_f90122ecffad.slice/crio-b209ab4fa67dfa5911dcdd2f26ac2e28c4361488b369e02c0c22069609e57ca9 WatchSource:0}: Error finding container b209ab4fa67dfa5911dcdd2f26ac2e28c4361488b369e02c0c22069609e57ca9: Status 404 returned error can't find the container with id b209ab4fa67dfa5911dcdd2f26ac2e28c4361488b369e02c0c22069609e57ca9 Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.648375 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:06:59 crc kubenswrapper[5018]: W1014 07:06:59.650983 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdbccee9_3a08_4d1b_9358_aa74613fa4d4.slice/crio-663b365e2061a29b0bf15587dd9f70220de7de7f4fa6ab623efb3f360b600bbc WatchSource:0}: Error finding container 663b365e2061a29b0bf15587dd9f70220de7de7f4fa6ab623efb3f360b600bbc: Status 404 returned error can't find the container with id 663b365e2061a29b0bf15587dd9f70220de7de7f4fa6ab623efb3f360b600bbc Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.721068 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.722370 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.728560 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.729948 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.730148 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.730314 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.730504 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.730725 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zqd7x" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.730909 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.737543 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910129 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910210 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910263 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910280 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910318 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910443 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9xx\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910489 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910660 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910724 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910826 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.910855 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.981999 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.984389 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.986422 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2qhgj" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.986589 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.986757 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.987029 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.987213 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.987284 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 07:06:59 crc kubenswrapper[5018]: I1014 07:06:59.987297 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.007342 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013096 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013147 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013198 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013228 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013274 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013301 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013350 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013369 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013397 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013422 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9xx\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.013444 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.014233 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.014498 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.014850 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.015134 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.015353 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.020785 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.021457 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.024097 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.024153 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.024545 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.032148 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9xx\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.046814 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " pod="openstack/rabbitmq-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.098916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" event={"ID":"b3255895-3e18-419b-a8fc-f90122ecffad","Type":"ContainerStarted","Data":"b209ab4fa67dfa5911dcdd2f26ac2e28c4361488b369e02c0c22069609e57ca9"} Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.100964 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" event={"ID":"cdbccee9-3a08-4d1b-9358-aa74613fa4d4","Type":"ContainerStarted","Data":"663b365e2061a29b0bf15587dd9f70220de7de7f4fa6ab623efb3f360b600bbc"} Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118269 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118389 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118425 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118453 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118467 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118534 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118583 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8f69\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118648 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.118729 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.221685 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8f69\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.221844 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.221874 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.221919 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222148 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222223 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222242 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222257 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222298 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222324 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.222720 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.223700 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.224206 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.224383 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.226816 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.227471 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.228120 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.233524 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.233693 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.235331 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8f69\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.252837 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.259255 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.309086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:07:00 crc kubenswrapper[5018]: I1014 07:07:00.348094 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.366930 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.373219 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.376491 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.376780 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.376900 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.377948 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.378568 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.378868 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-pvcll" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.384876 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474634 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnpkb\" (UniqueName: \"kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474681 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474720 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474740 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474787 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474841 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474867 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.474904 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576459 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576526 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576564 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576603 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnpkb\" (UniqueName: \"kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576647 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576685 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576721 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576738 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.576810 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.577973 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.578378 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.579054 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.579350 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.579776 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.582315 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.583034 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.599102 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.601670 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.602021 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnpkb\" (UniqueName: \"kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb\") pod \"openstack-galera-0\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " pod="openstack/openstack-galera-0" Oct 14 07:07:01 crc kubenswrapper[5018]: I1014 07:07:01.705605 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.718643 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.720227 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.722100 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.724010 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.724219 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.725091 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9p4tx" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.727519 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794116 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794196 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794240 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794264 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794282 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794306 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794326 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sflj7\" (UniqueName: \"kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794348 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.794364 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895550 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895646 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895681 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895855 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895900 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895924 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.895939 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.897497 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.897518 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sflj7\" (UniqueName: \"kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.896916 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.897441 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.896044 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.896350 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.899502 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.902401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.903439 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.912880 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.916964 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sflj7\" (UniqueName: \"kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:02 crc kubenswrapper[5018]: I1014 07:07:02.921329 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.040294 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.252747 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.254353 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.257519 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.260926 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qjs7t" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.261246 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.276112 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.411641 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.411865 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.411903 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq8dc\" (UniqueName: \"kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.412093 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.412169 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.518189 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.518251 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq8dc\" (UniqueName: \"kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.518315 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.518347 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.518371 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.519277 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.520784 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.522827 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.536384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.541154 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq8dc\" (UniqueName: \"kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc\") pod \"memcached-0\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " pod="openstack/memcached-0" Oct 14 07:07:03 crc kubenswrapper[5018]: I1014 07:07:03.574235 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.767660 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.769175 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.771254 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-q9wml" Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.775708 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.841812 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqpfg\" (UniqueName: \"kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg\") pod \"kube-state-metrics-0\" (UID: \"25658566-e65b-48fe-aa02-1df2ab82b526\") " pod="openstack/kube-state-metrics-0" Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.945362 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqpfg\" (UniqueName: \"kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg\") pod \"kube-state-metrics-0\" (UID: \"25658566-e65b-48fe-aa02-1df2ab82b526\") " pod="openstack/kube-state-metrics-0" Oct 14 07:07:04 crc kubenswrapper[5018]: I1014 07:07:04.987837 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqpfg\" (UniqueName: \"kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg\") pod \"kube-state-metrics-0\" (UID: \"25658566-e65b-48fe-aa02-1df2ab82b526\") " pod="openstack/kube-state-metrics-0" Oct 14 07:07:05 crc kubenswrapper[5018]: I1014 07:07:05.094340 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:07:07 crc kubenswrapper[5018]: I1014 07:07:07.361736 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.650296 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.652565 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.655317 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.655823 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.655837 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tjk89" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.665791 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.671795 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.673973 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.695603 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.805077 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.806439 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.808908 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.811877 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.812007 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.811877 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k5ptb" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.812266 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.817397 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.817870 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.817963 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.818034 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.827747 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.827966 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.828056 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.828179 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9gd9\" (UniqueName: \"kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.828298 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.828399 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frrj\" (UniqueName: \"kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.831681 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.831802 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.831931 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.832056 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.934121 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.936692 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.936860 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.937695 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqsx9\" (UniqueName: \"kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.937093 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.937820 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.937945 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938011 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938081 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9gd9\" (UniqueName: \"kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938146 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938180 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frrj\" (UniqueName: \"kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938202 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938245 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938317 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938343 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938373 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938401 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938452 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938479 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938506 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.938527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939194 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939336 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939506 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939690 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939740 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939779 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.939965 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.941573 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.943911 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.958804 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frrj\" (UniqueName: \"kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.958890 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9gd9\" (UniqueName: \"kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9\") pod \"ovn-controller-ovs-52nkw\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.959946 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs\") pod \"ovn-controller-gs786\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.976328 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786" Oct 14 07:07:09 crc kubenswrapper[5018]: I1014 07:07:09.999041 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040758 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040818 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqsx9\" (UniqueName: \"kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040846 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040866 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040926 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040942 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040961 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.040990 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.042065 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.042542 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.042566 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.042574 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.048678 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.049398 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.056819 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqsx9\" (UniqueName: \"kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.064299 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.065182 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:10 crc kubenswrapper[5018]: I1014 07:07:10.121454 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.101812 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.103257 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.112088 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.112111 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.112226 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vm4fc" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.112232 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.124392 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.260635 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.260693 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.260726 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.260908 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.260973 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.261059 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.261115 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzb74\" (UniqueName: \"kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.261199 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362160 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362213 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362262 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362289 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362313 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzb74\" (UniqueName: \"kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362367 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362397 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.362768 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.363396 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.365591 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.366794 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.370832 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.375354 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.375810 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.382299 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzb74\" (UniqueName: \"kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.383021 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.436518 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:11 crc kubenswrapper[5018]: W1014 07:07:11.631689 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ddcde8d_45fd_419b_8751_18db89bf729d.slice/crio-f348120b2e00e1e900a787ef0d2a2e30b118ad5602a3c48534dbce41a0172cc7 WatchSource:0}: Error finding container f348120b2e00e1e900a787ef0d2a2e30b118ad5602a3c48534dbce41a0172cc7: Status 404 returned error can't find the container with id f348120b2e00e1e900a787ef0d2a2e30b118ad5602a3c48534dbce41a0172cc7 Oct 14 07:07:11 crc kubenswrapper[5018]: I1014 07:07:11.635204 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:07:12 crc kubenswrapper[5018]: I1014 07:07:12.210644 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerStarted","Data":"f348120b2e00e1e900a787ef0d2a2e30b118ad5602a3c48534dbce41a0172cc7"} Oct 14 07:07:13 crc kubenswrapper[5018]: E1014 07:07:13.940895 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:07:13 crc kubenswrapper[5018]: E1014 07:07:13.941587 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fql6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6948694bd9-8hs6c_openstack(e3da82b5-168a-4504-b171-6d44e63f4d3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:07:13 crc kubenswrapper[5018]: E1014 07:07:13.942723 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" podUID="e3da82b5-168a-4504-b171-6d44e63f4d3b" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.043274 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.043435 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6kwp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-86f694bf-l2hnw_openstack(b3255895-3e18-419b-a8fc-f90122ecffad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.044811 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.086844 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.087014 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mb79m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d487d97d7-9xpdm_openstack(331cc27c-304b-4264-a43f-4a03657c4435): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:07:14 crc kubenswrapper[5018]: E1014 07:07:14.089217 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" podUID="331cc27c-304b-4264-a43f-4a03657c4435" Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.441843 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.471952 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.479556 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.543848 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: W1014 07:07:14.622406 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbaa6d249_3441_4d16_9a74_a5232861aa21.slice/crio-4c9a3e52f586edf8c75eb084c5173d36c670a666390cc158e3ece7447c634a48 WatchSource:0}: Error finding container 4c9a3e52f586edf8c75eb084c5173d36c670a666390cc158e3ece7447c634a48: Status 404 returned error can't find the container with id 4c9a3e52f586edf8c75eb084c5173d36c670a666390cc158e3ece7447c634a48 Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.785988 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.798062 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.914311 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:07:14 crc kubenswrapper[5018]: W1014 07:07:14.920776 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1751d1e2_6a95_4a5f_b588_a58bffb94896.slice/crio-1e17029dd8449e4c33b88e3367a079228a62bf3acbb37159be215d24794146e3 WatchSource:0}: Error finding container 1e17029dd8449e4c33b88e3367a079228a62bf3acbb37159be215d24794146e3: Status 404 returned error can't find the container with id 1e17029dd8449e4c33b88e3367a079228a62bf3acbb37159be215d24794146e3 Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.978008 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:07:14 crc kubenswrapper[5018]: I1014 07:07:14.983852 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.002088 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:07:15 crc kubenswrapper[5018]: W1014 07:07:15.007897 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2372cde_de88_46c2_abc8_0bd1a9e7b961.slice/crio-c453ceed07ab2e563923f41d67fe3f669d93551cadb48fc652de1444a662220e WatchSource:0}: Error finding container c453ceed07ab2e563923f41d67fe3f669d93551cadb48fc652de1444a662220e: Status 404 returned error can't find the container with id c453ceed07ab2e563923f41d67fe3f669d93551cadb48fc652de1444a662220e Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.149683 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb79m\" (UniqueName: \"kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m\") pod \"331cc27c-304b-4264-a43f-4a03657c4435\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.149776 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config\") pod \"e3da82b5-168a-4504-b171-6d44e63f4d3b\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.149884 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config\") pod \"331cc27c-304b-4264-a43f-4a03657c4435\" (UID: \"331cc27c-304b-4264-a43f-4a03657c4435\") " Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.149941 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc\") pod \"e3da82b5-168a-4504-b171-6d44e63f4d3b\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.149982 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fql6b\" (UniqueName: \"kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b\") pod \"e3da82b5-168a-4504-b171-6d44e63f4d3b\" (UID: \"e3da82b5-168a-4504-b171-6d44e63f4d3b\") " Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.150332 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config" (OuterVolumeSpecName: "config") pod "e3da82b5-168a-4504-b171-6d44e63f4d3b" (UID: "e3da82b5-168a-4504-b171-6d44e63f4d3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.150381 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config" (OuterVolumeSpecName: "config") pod "331cc27c-304b-4264-a43f-4a03657c4435" (UID: "331cc27c-304b-4264-a43f-4a03657c4435"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.150767 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3da82b5-168a-4504-b171-6d44e63f4d3b" (UID: "e3da82b5-168a-4504-b171-6d44e63f4d3b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.164271 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b" (OuterVolumeSpecName: "kube-api-access-fql6b") pod "e3da82b5-168a-4504-b171-6d44e63f4d3b" (UID: "e3da82b5-168a-4504-b171-6d44e63f4d3b"). InnerVolumeSpecName "kube-api-access-fql6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.171871 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m" (OuterVolumeSpecName: "kube-api-access-mb79m") pod "331cc27c-304b-4264-a43f-4a03657c4435" (UID: "331cc27c-304b-4264-a43f-4a03657c4435"). InnerVolumeSpecName "kube-api-access-mb79m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.241258 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786" event={"ID":"ecd8d742-9217-4dd6-a035-49fab332a574","Type":"ContainerStarted","Data":"6b35061bda013c035c4d6f3e6ae981f891709a14cd9e86badf29fade130b68b1"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.242716 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" event={"ID":"e3da82b5-168a-4504-b171-6d44e63f4d3b","Type":"ContainerDied","Data":"351974fd17b1d860a44337f00ac7d52e365d93a88aaac717d5a29e2c0f12f1b1"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.242807 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-8hs6c" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.247611 5018 generic.go:334] "Generic (PLEG): container finished" podID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerID="3375d853d1bdfe85265294e950118e19f636fb1c462206347e6339a1a6c6b702" exitCode=0 Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.248017 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" event={"ID":"cdbccee9-3a08-4d1b-9358-aa74613fa4d4","Type":"ContainerDied","Data":"3375d853d1bdfe85265294e950118e19f636fb1c462206347e6339a1a6c6b702"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.252023 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e","Type":"ContainerStarted","Data":"80330bc7a1cf897e49fc60598b544559c0f8ffe3226288e1802790d454d95900"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254784 5018 generic.go:334] "Generic (PLEG): container finished" podID="b3255895-3e18-419b-a8fc-f90122ecffad" containerID="97944c41abb882ce6506402ba39f106fc9a0cd59b9ccfb08afd9905eddc184ac" exitCode=0 Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254812 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" event={"ID":"b3255895-3e18-419b-a8fc-f90122ecffad","Type":"ContainerDied","Data":"97944c41abb882ce6506402ba39f106fc9a0cd59b9ccfb08afd9905eddc184ac"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254793 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fql6b\" (UniqueName: \"kubernetes.io/projected/e3da82b5-168a-4504-b171-6d44e63f4d3b-kube-api-access-fql6b\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254916 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb79m\" (UniqueName: \"kubernetes.io/projected/331cc27c-304b-4264-a43f-4a03657c4435-kube-api-access-mb79m\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254933 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254962 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331cc27c-304b-4264-a43f-4a03657c4435-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.254973 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3da82b5-168a-4504-b171-6d44e63f4d3b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.278656 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25658566-e65b-48fe-aa02-1df2ab82b526","Type":"ContainerStarted","Data":"9716a1206a89523fa2008c6f7ef3acd6907d2afe88e152b28ff8ef2c8a27e5aa"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.280115 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerStarted","Data":"c453ceed07ab2e563923f41d67fe3f669d93551cadb48fc652de1444a662220e"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.281245 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerStarted","Data":"4c9a3e52f586edf8c75eb084c5173d36c670a666390cc158e3ece7447c634a48"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.284903 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerStarted","Data":"1e17029dd8449e4c33b88e3367a079228a62bf3acbb37159be215d24794146e3"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.287709 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.287849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-9xpdm" event={"ID":"331cc27c-304b-4264-a43f-4a03657c4435","Type":"ContainerDied","Data":"2f31c88923ee234e057545479a668dfc61fec238889f712707a86c15339a3712"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.298823 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerStarted","Data":"a6d658ec166fe9a0201d6b6a684b09e0b7e1e06c96418a5221797ee370089267"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.300595 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerStarted","Data":"d63779d331926f3a753bff10e4496ef456d30c4baa486cde1f2edcf14ed6d9ae"} Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.337058 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.353227 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-8hs6c"] Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.375433 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.378712 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-9xpdm"] Oct 14 07:07:15 crc kubenswrapper[5018]: I1014 07:07:15.749021 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.313123 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" event={"ID":"cdbccee9-3a08-4d1b-9358-aa74613fa4d4","Type":"ContainerStarted","Data":"877586f8c9e415608988983abdb69010b77342ae28cdb031c879166b820af0af"} Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.313642 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.316211 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" event={"ID":"b3255895-3e18-419b-a8fc-f90122ecffad","Type":"ContainerStarted","Data":"3e3f82a70f1b7f2ec8b5b8ed040a7c31fb5fc27fd92e849c768857aa08d3b5e8"} Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.316501 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.328747 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" podStartSLOduration=3.910770314 podStartE2EDuration="18.328735276s" podCreationTimestamp="2025-10-14 07:06:58 +0000 UTC" firstStartedPulling="2025-10-14 07:06:59.653001419 +0000 UTC m=+1036.237048046" lastFinishedPulling="2025-10-14 07:07:14.070966381 +0000 UTC m=+1050.655013008" observedRunningTime="2025-10-14 07:07:16.328195051 +0000 UTC m=+1052.912241678" watchObservedRunningTime="2025-10-14 07:07:16.328735276 +0000 UTC m=+1052.912781903" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.343361 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" podStartSLOduration=-9223372018.511429 podStartE2EDuration="18.343346178s" podCreationTimestamp="2025-10-14 07:06:58 +0000 UTC" firstStartedPulling="2025-10-14 07:06:59.223198214 +0000 UTC m=+1035.807244841" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:07:16.342527205 +0000 UTC m=+1052.926573832" watchObservedRunningTime="2025-10-14 07:07:16.343346178 +0000 UTC m=+1052.927392805" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.614019 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331cc27c-304b-4264-a43f-4a03657c4435" path="/var/lib/kubelet/pods/331cc27c-304b-4264-a43f-4a03657c4435/volumes" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.614372 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3da82b5-168a-4504-b171-6d44e63f4d3b" path="/var/lib/kubelet/pods/e3da82b5-168a-4504-b171-6d44e63f4d3b/volumes" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.769775 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.772764 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.776049 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.779587 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.885008 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899637 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899694 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899742 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899801 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xphpj\" (UniqueName: \"kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899827 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.899867 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.901776 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.904898 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.909579 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 07:07:16 crc kubenswrapper[5018]: I1014 07:07:16.921138 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001388 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001455 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xphpj\" (UniqueName: \"kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001487 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001529 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001567 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001585 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.001857 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.002825 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.003422 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.007557 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.009792 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.024286 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xphpj\" (UniqueName: \"kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj\") pod \"ovn-controller-metrics-7pr99\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.063445 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.086508 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.087716 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.089672 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.096265 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.101058 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.102944 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk5jr\" (UniqueName: \"kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.103024 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.103058 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.103089 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.205035 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk5jr\" (UniqueName: \"kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.205189 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.205239 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206092 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206261 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206320 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206371 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrzzp\" (UniqueName: \"kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.206527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.207410 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.207496 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.207847 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.228458 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk5jr\" (UniqueName: \"kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr\") pod \"dnsmasq-dns-57c5bcf59c-5nv7x\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.228914 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.307874 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.307966 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.307998 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrzzp\" (UniqueName: \"kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.308036 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.308098 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.309166 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.309403 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.309856 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.310161 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.398881 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrzzp\" (UniqueName: \"kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp\") pod \"dnsmasq-dns-d49c4d845-m7xc5\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:17 crc kubenswrapper[5018]: I1014 07:07:17.417681 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:18 crc kubenswrapper[5018]: I1014 07:07:18.340213 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="dnsmasq-dns" containerID="cri-o://3e3f82a70f1b7f2ec8b5b8ed040a7c31fb5fc27fd92e849c768857aa08d3b5e8" gracePeriod=10 Oct 14 07:07:18 crc kubenswrapper[5018]: I1014 07:07:18.340327 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="dnsmasq-dns" containerID="cri-o://877586f8c9e415608988983abdb69010b77342ae28cdb031c879166b820af0af" gracePeriod=10 Oct 14 07:07:20 crc kubenswrapper[5018]: W1014 07:07:20.301261 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17e34896_a3ff_4810_a21e_b8c5ccad8379.slice/crio-2814725534bf0457611cad01eed2a88b8298cab45bb48662d892ff43122e2050 WatchSource:0}: Error finding container 2814725534bf0457611cad01eed2a88b8298cab45bb48662d892ff43122e2050: Status 404 returned error can't find the container with id 2814725534bf0457611cad01eed2a88b8298cab45bb48662d892ff43122e2050 Oct 14 07:07:20 crc kubenswrapper[5018]: I1014 07:07:20.361268 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerStarted","Data":"2814725534bf0457611cad01eed2a88b8298cab45bb48662d892ff43122e2050"} Oct 14 07:07:20 crc kubenswrapper[5018]: I1014 07:07:20.367037 5018 generic.go:334] "Generic (PLEG): container finished" podID="b3255895-3e18-419b-a8fc-f90122ecffad" containerID="3e3f82a70f1b7f2ec8b5b8ed040a7c31fb5fc27fd92e849c768857aa08d3b5e8" exitCode=0 Oct 14 07:07:20 crc kubenswrapper[5018]: I1014 07:07:20.367104 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" event={"ID":"b3255895-3e18-419b-a8fc-f90122ecffad","Type":"ContainerDied","Data":"3e3f82a70f1b7f2ec8b5b8ed040a7c31fb5fc27fd92e849c768857aa08d3b5e8"} Oct 14 07:07:20 crc kubenswrapper[5018]: I1014 07:07:20.369803 5018 generic.go:334] "Generic (PLEG): container finished" podID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerID="877586f8c9e415608988983abdb69010b77342ae28cdb031c879166b820af0af" exitCode=0 Oct 14 07:07:20 crc kubenswrapper[5018]: I1014 07:07:20.369847 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" event={"ID":"cdbccee9-3a08-4d1b-9358-aa74613fa4d4","Type":"ContainerDied","Data":"877586f8c9e415608988983abdb69010b77342ae28cdb031c879166b820af0af"} Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.888374 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.896012 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.964759 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc\") pod \"b3255895-3e18-419b-a8fc-f90122ecffad\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.964978 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6kwp\" (UniqueName: \"kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp\") pod \"b3255895-3e18-419b-a8fc-f90122ecffad\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.965044 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config\") pod \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.965097 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config\") pod \"b3255895-3e18-419b-a8fc-f90122ecffad\" (UID: \"b3255895-3e18-419b-a8fc-f90122ecffad\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.965148 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc\") pod \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.965198 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8krps\" (UniqueName: \"kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps\") pod \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\" (UID: \"cdbccee9-3a08-4d1b-9358-aa74613fa4d4\") " Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.971206 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps" (OuterVolumeSpecName: "kube-api-access-8krps") pod "cdbccee9-3a08-4d1b-9358-aa74613fa4d4" (UID: "cdbccee9-3a08-4d1b-9358-aa74613fa4d4"). InnerVolumeSpecName "kube-api-access-8krps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:24 crc kubenswrapper[5018]: I1014 07:07:24.983544 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp" (OuterVolumeSpecName: "kube-api-access-n6kwp") pod "b3255895-3e18-419b-a8fc-f90122ecffad" (UID: "b3255895-3e18-419b-a8fc-f90122ecffad"). InnerVolumeSpecName "kube-api-access-n6kwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.012572 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config" (OuterVolumeSpecName: "config") pod "b3255895-3e18-419b-a8fc-f90122ecffad" (UID: "b3255895-3e18-419b-a8fc-f90122ecffad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.018118 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3255895-3e18-419b-a8fc-f90122ecffad" (UID: "b3255895-3e18-419b-a8fc-f90122ecffad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.021630 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdbccee9-3a08-4d1b-9358-aa74613fa4d4" (UID: "cdbccee9-3a08-4d1b-9358-aa74613fa4d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.024234 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config" (OuterVolumeSpecName: "config") pod "cdbccee9-3a08-4d1b-9358-aa74613fa4d4" (UID: "cdbccee9-3a08-4d1b-9358-aa74613fa4d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067394 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067422 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6kwp\" (UniqueName: \"kubernetes.io/projected/b3255895-3e18-419b-a8fc-f90122ecffad-kube-api-access-n6kwp\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067436 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067448 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3255895-3e18-419b-a8fc-f90122ecffad-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067461 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.067470 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8krps\" (UniqueName: \"kubernetes.io/projected/cdbccee9-3a08-4d1b-9358-aa74613fa4d4-kube-api-access-8krps\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.432591 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" event={"ID":"cdbccee9-3a08-4d1b-9358-aa74613fa4d4","Type":"ContainerDied","Data":"663b365e2061a29b0bf15587dd9f70220de7de7f4fa6ab623efb3f360b600bbc"} Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.432655 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.432716 5018 scope.go:117] "RemoveContainer" containerID="877586f8c9e415608988983abdb69010b77342ae28cdb031c879166b820af0af" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.434421 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" event={"ID":"b3255895-3e18-419b-a8fc-f90122ecffad","Type":"ContainerDied","Data":"b209ab4fa67dfa5911dcdd2f26ac2e28c4361488b369e02c0c22069609e57ca9"} Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.434486 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.468591 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.474975 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-l2hnw"] Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.484120 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.491747 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-fwmld"] Oct 14 07:07:25 crc kubenswrapper[5018]: I1014 07:07:25.773382 5018 scope.go:117] "RemoveContainer" containerID="3375d853d1bdfe85265294e950118e19f636fb1c462206347e6339a1a6c6b702" Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.088232 5018 scope.go:117] "RemoveContainer" containerID="3e3f82a70f1b7f2ec8b5b8ed040a7c31fb5fc27fd92e849c768857aa08d3b5e8" Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.188502 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.197064 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.293242 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:26 crc kubenswrapper[5018]: E1014 07:07:26.508731 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 14 07:07:26 crc kubenswrapper[5018]: E1014 07:07:26.508767 5018 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 14 07:07:26 crc kubenswrapper[5018]: E1014 07:07:26.508880 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zqpfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(25658566-e65b-48fe-aa02-1df2ab82b526): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:07:26 crc kubenswrapper[5018]: E1014 07:07:26.509943 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" Oct 14 07:07:26 crc kubenswrapper[5018]: W1014 07:07:26.525144 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b467eaf_55a9_48a2_a852_008b6ebca0aa.slice/crio-f8f148db13462c40e71767d224087f820ee1835089b9ab67989cd9b16850a8fa WatchSource:0}: Error finding container f8f148db13462c40e71767d224087f820ee1835089b9ab67989cd9b16850a8fa: Status 404 returned error can't find the container with id f8f148db13462c40e71767d224087f820ee1835089b9ab67989cd9b16850a8fa Oct 14 07:07:26 crc kubenswrapper[5018]: W1014 07:07:26.527256 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9dc8f93_29e8_482a_9a26_418541db4e61.slice/crio-ec522f3017902d9e0ede06a300a45929d0fe3c6b65b874a2f053f52ad45f61f4 WatchSource:0}: Error finding container ec522f3017902d9e0ede06a300a45929d0fe3c6b65b874a2f053f52ad45f61f4: Status 404 returned error can't find the container with id ec522f3017902d9e0ede06a300a45929d0fe3c6b65b874a2f053f52ad45f61f4 Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.549849 5018 scope.go:117] "RemoveContainer" containerID="97944c41abb882ce6506402ba39f106fc9a0cd59b9ccfb08afd9905eddc184ac" Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.613752 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" path="/var/lib/kubelet/pods/b3255895-3e18-419b-a8fc-f90122ecffad/volumes" Oct 14 07:07:26 crc kubenswrapper[5018]: I1014 07:07:26.614579 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" path="/var/lib/kubelet/pods/cdbccee9-3a08-4d1b-9358-aa74613fa4d4/volumes" Oct 14 07:07:27 crc kubenswrapper[5018]: I1014 07:07:27.458439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7pr99" event={"ID":"0f737b49-f2ba-4a7c-9b07-d934d5122363","Type":"ContainerStarted","Data":"1a8e95b24b6768632df6c1d84ae1755081180d78a7737d67ae8782e519da4518"} Oct 14 07:07:27 crc kubenswrapper[5018]: I1014 07:07:27.462437 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" event={"ID":"e9dc8f93-29e8-482a-9a26-418541db4e61","Type":"ContainerStarted","Data":"ec522f3017902d9e0ede06a300a45929d0fe3c6b65b874a2f053f52ad45f61f4"} Oct 14 07:07:27 crc kubenswrapper[5018]: I1014 07:07:27.464049 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerStarted","Data":"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be"} Oct 14 07:07:27 crc kubenswrapper[5018]: I1014 07:07:27.466589 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" event={"ID":"1b467eaf-55a9-48a2-a852-008b6ebca0aa","Type":"ContainerStarted","Data":"f8f148db13462c40e71767d224087f820ee1835089b9ab67989cd9b16850a8fa"} Oct 14 07:07:27 crc kubenswrapper[5018]: E1014 07:07:27.468178 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.479350 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerStarted","Data":"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.484274 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786" event={"ID":"ecd8d742-9217-4dd6-a035-49fab332a574","Type":"ContainerStarted","Data":"76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.484340 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gs786" Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.498820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerStarted","Data":"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.502296 5018 generic.go:334] "Generic (PLEG): container finished" podID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerID="2c75064d547e5956672b603f9c1a06a4e5637ac16ca15b6e701e1c42346cfa93" exitCode=0 Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.502372 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerDied","Data":"2c75064d547e5956672b603f9c1a06a4e5637ac16ca15b6e701e1c42346cfa93"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.505709 5018 generic.go:334] "Generic (PLEG): container finished" podID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerID="8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767" exitCode=0 Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.505778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" event={"ID":"e9dc8f93-29e8-482a-9a26-418541db4e61","Type":"ContainerDied","Data":"8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.510536 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerStarted","Data":"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.512817 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gs786" podStartSLOduration=8.248588001 podStartE2EDuration="19.51280335s" podCreationTimestamp="2025-10-14 07:07:09 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.804536249 +0000 UTC m=+1051.388582876" lastFinishedPulling="2025-10-14 07:07:26.068751598 +0000 UTC m=+1062.652798225" observedRunningTime="2025-10-14 07:07:28.5088234 +0000 UTC m=+1065.092870167" watchObservedRunningTime="2025-10-14 07:07:28.51280335 +0000 UTC m=+1065.096849977" Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.513739 5018 generic.go:334] "Generic (PLEG): container finished" podID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerID="f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a" exitCode=0 Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.513836 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" event={"ID":"1b467eaf-55a9-48a2-a852-008b6ebca0aa","Type":"ContainerDied","Data":"f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.515685 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerStarted","Data":"124f8aee2f1028a966a6895f7aa8cc6cea30f99e63669865c80d68ed595f9d7f"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.517439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e","Type":"ContainerStarted","Data":"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.517516 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.519811 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerStarted","Data":"f1c52505d00c71d74b81b018f8807279db186987ee6fc9cf83ce290b691b50e9"} Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.624938 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.468828632 podStartE2EDuration="25.624918786s" podCreationTimestamp="2025-10-14 07:07:03 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.459960856 +0000 UTC m=+1051.044007483" lastFinishedPulling="2025-10-14 07:07:25.616051 +0000 UTC m=+1062.200097637" observedRunningTime="2025-10-14 07:07:28.621921663 +0000 UTC m=+1065.205968290" watchObservedRunningTime="2025-10-14 07:07:28.624918786 +0000 UTC m=+1065.208965413" Oct 14 07:07:28 crc kubenswrapper[5018]: I1014 07:07:28.923700 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86f694bf-l2hnw" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.100:5353: i/o timeout" Oct 14 07:07:29 crc kubenswrapper[5018]: I1014 07:07:29.191734 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7869c47d6c-fwmld" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.101:5353: i/o timeout" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.535058 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerStarted","Data":"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.538286 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" event={"ID":"1b467eaf-55a9-48a2-a852-008b6ebca0aa","Type":"ContainerStarted","Data":"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.538419 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.541572 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7pr99" event={"ID":"0f737b49-f2ba-4a7c-9b07-d934d5122363","Type":"ContainerStarted","Data":"3b16b9017596763bb0f70d7dd6f8aca7f1f4fbb0b4ed9d4a7005193cc2610d01"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.545932 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerStarted","Data":"015528ec03a9ccae5d8175782d21df263addf1920d63d59b5c8dd6f6c5ecbd26"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.548295 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerStarted","Data":"312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.550955 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" event={"ID":"e9dc8f93-29e8-482a-9a26-418541db4e61","Type":"ContainerStarted","Data":"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3"} Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.551229 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.565989 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.456688737 podStartE2EDuration="22.565969794s" podCreationTimestamp="2025-10-14 07:07:08 +0000 UTC" firstStartedPulling="2025-10-14 07:07:15.010092206 +0000 UTC m=+1051.594138833" lastFinishedPulling="2025-10-14 07:07:30.119373263 +0000 UTC m=+1066.703419890" observedRunningTime="2025-10-14 07:07:30.55890645 +0000 UTC m=+1067.142953077" watchObservedRunningTime="2025-10-14 07:07:30.565969794 +0000 UTC m=+1067.150016431" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.597454 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.212436525 podStartE2EDuration="20.59742833s" podCreationTimestamp="2025-10-14 07:07:10 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.922692741 +0000 UTC m=+1051.506739358" lastFinishedPulling="2025-10-14 07:07:30.307684536 +0000 UTC m=+1066.891731163" observedRunningTime="2025-10-14 07:07:30.583141757 +0000 UTC m=+1067.167188384" watchObservedRunningTime="2025-10-14 07:07:30.59742833 +0000 UTC m=+1067.181474977" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.606989 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" podStartSLOduration=13.606967063 podStartE2EDuration="13.606967063s" podCreationTimestamp="2025-10-14 07:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:07:30.602316685 +0000 UTC m=+1067.186363312" watchObservedRunningTime="2025-10-14 07:07:30.606967063 +0000 UTC m=+1067.191013710" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.624839 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" podStartSLOduration=14.624821714 podStartE2EDuration="14.624821714s" podCreationTimestamp="2025-10-14 07:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:07:30.618426048 +0000 UTC m=+1067.202472675" watchObservedRunningTime="2025-10-14 07:07:30.624821714 +0000 UTC m=+1067.208868341" Oct 14 07:07:30 crc kubenswrapper[5018]: I1014 07:07:30.646431 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7pr99" podStartSLOduration=11.018346982 podStartE2EDuration="14.646411118s" podCreationTimestamp="2025-10-14 07:07:16 +0000 UTC" firstStartedPulling="2025-10-14 07:07:26.517977181 +0000 UTC m=+1063.102023808" lastFinishedPulling="2025-10-14 07:07:30.146041317 +0000 UTC m=+1066.730087944" observedRunningTime="2025-10-14 07:07:30.63522972 +0000 UTC m=+1067.219276357" watchObservedRunningTime="2025-10-14 07:07:30.646411118 +0000 UTC m=+1067.230457755" Oct 14 07:07:30 crc kubenswrapper[5018]: E1014 07:07:30.894518 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ca7068c_e306_467d_972d_6ec8439d2b85.slice/crio-92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ca7068c_e306_467d_972d_6ec8439d2b85.slice/crio-conmon-92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.122313 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.188275 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.437397 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.560279 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerStarted","Data":"7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4"} Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.560647 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.560779 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.561825 5018 generic.go:334] "Generic (PLEG): container finished" podID="baa6d249-3441-4d16-9a74-a5232861aa21" containerID="3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834" exitCode=0 Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.561889 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerDied","Data":"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834"} Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.563525 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerID="92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be" exitCode=0 Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.563561 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerDied","Data":"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be"} Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.564076 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:31 crc kubenswrapper[5018]: I1014 07:07:31.638800 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-52nkw" podStartSLOduration=17.173794888 podStartE2EDuration="22.638780239s" podCreationTimestamp="2025-10-14 07:07:09 +0000 UTC" firstStartedPulling="2025-10-14 07:07:20.307499914 +0000 UTC m=+1056.891546551" lastFinishedPulling="2025-10-14 07:07:25.772485255 +0000 UTC m=+1062.356531902" observedRunningTime="2025-10-14 07:07:31.597410191 +0000 UTC m=+1068.181456848" watchObservedRunningTime="2025-10-14 07:07:31.638780239 +0000 UTC m=+1068.222826866" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.437270 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.512993 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.578690 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerStarted","Data":"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e"} Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.581872 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerStarted","Data":"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897"} Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.615397 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.17212836 podStartE2EDuration="31.615364826s" podCreationTimestamp="2025-10-14 07:07:01 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.625764689 +0000 UTC m=+1051.209811306" lastFinishedPulling="2025-10-14 07:07:26.069001155 +0000 UTC m=+1062.653047772" observedRunningTime="2025-10-14 07:07:32.607404987 +0000 UTC m=+1069.191451654" watchObservedRunningTime="2025-10-14 07:07:32.615364826 +0000 UTC m=+1069.199411493" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.648381 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.516980371 podStartE2EDuration="32.648351524s" podCreationTimestamp="2025-10-14 07:07:00 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.493774317 +0000 UTC m=+1051.077820944" lastFinishedPulling="2025-10-14 07:07:25.62514546 +0000 UTC m=+1062.209192097" observedRunningTime="2025-10-14 07:07:32.63222476 +0000 UTC m=+1069.216271397" watchObservedRunningTime="2025-10-14 07:07:32.648351524 +0000 UTC m=+1069.232398181" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.662430 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.663688 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.942659 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:07:32 crc kubenswrapper[5018]: E1014 07:07:32.942990 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="init" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943005 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="init" Oct 14 07:07:32 crc kubenswrapper[5018]: E1014 07:07:32.943019 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943028 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: E1014 07:07:32.943037 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943043 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: E1014 07:07:32.943059 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="init" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943065 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="init" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943212 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdbccee9-3a08-4d1b-9358-aa74613fa4d4" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.943223 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3255895-3e18-419b-a8fc-f90122ecffad" containerName="dnsmasq-dns" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.944106 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.947180 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.947237 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.947967 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.952368 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pltgn" Oct 14 07:07:32 crc kubenswrapper[5018]: I1014 07:07:32.965505 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.035984 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr476\" (UniqueName: \"kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036028 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036084 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036113 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036330 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036498 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.036551 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.042699 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.042746 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144741 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144848 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144894 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144913 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144950 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr476\" (UniqueName: \"kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.144978 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.145041 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.145601 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.146096 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.146893 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.149533 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.158408 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.170206 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.176989 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr476\" (UniqueName: \"kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476\") pod \"ovn-northd-0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.266683 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.576909 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 07:07:33 crc kubenswrapper[5018]: I1014 07:07:33.689925 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:07:34 crc kubenswrapper[5018]: I1014 07:07:34.600481 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerStarted","Data":"d121869711c7268cef9749d0de08bc8fc5607bbc0e7fd583dd9ac0b2cb48a58b"} Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.222949 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.223452 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="dnsmasq-dns" containerID="cri-o://7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1" gracePeriod=10 Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.230957 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.254383 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.258190 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.274813 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.286345 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.286424 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.286481 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.286519 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.286551 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnxr4\" (UniqueName: \"kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.390320 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.390377 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.391534 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.391657 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.391737 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnxr4\" (UniqueName: \"kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.391803 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.391882 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.392356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.392602 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.411325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnxr4\" (UniqueName: \"kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4\") pod \"dnsmasq-dns-c757dd68f-whbm8\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.590896 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.608892 5018 generic.go:334] "Generic (PLEG): container finished" podID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerID="7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1" exitCode=0 Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.608961 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" event={"ID":"1b467eaf-55a9-48a2-a852-008b6ebca0aa","Type":"ContainerDied","Data":"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1"} Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.608988 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" event={"ID":"1b467eaf-55a9-48a2-a852-008b6ebca0aa","Type":"ContainerDied","Data":"f8f148db13462c40e71767d224087f820ee1835089b9ab67989cd9b16850a8fa"} Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.609004 5018 scope.go:117] "RemoveContainer" containerID="7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.609096 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c5bcf59c-5nv7x" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.612360 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerStarted","Data":"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a"} Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.612397 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerStarted","Data":"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321"} Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.613082 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.617584 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.638819 5018 scope.go:117] "RemoveContainer" containerID="f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.643150 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.53806612 podStartE2EDuration="3.643128152s" podCreationTimestamp="2025-10-14 07:07:32 +0000 UTC" firstStartedPulling="2025-10-14 07:07:33.712364636 +0000 UTC m=+1070.296411263" lastFinishedPulling="2025-10-14 07:07:34.817426668 +0000 UTC m=+1071.401473295" observedRunningTime="2025-10-14 07:07:35.628725446 +0000 UTC m=+1072.212772073" watchObservedRunningTime="2025-10-14 07:07:35.643128152 +0000 UTC m=+1072.227174789" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.678281 5018 scope.go:117] "RemoveContainer" containerID="7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1" Oct 14 07:07:35 crc kubenswrapper[5018]: E1014 07:07:35.678689 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1\": container with ID starting with 7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1 not found: ID does not exist" containerID="7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.678722 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1"} err="failed to get container status \"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1\": rpc error: code = NotFound desc = could not find container \"7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1\": container with ID starting with 7f4e64759fd2df08328b03f0602440c10b4f4ecad050d9775e254ee502fe73c1 not found: ID does not exist" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.678745 5018 scope.go:117] "RemoveContainer" containerID="f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a" Oct 14 07:07:35 crc kubenswrapper[5018]: E1014 07:07:35.678980 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a\": container with ID starting with f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a not found: ID does not exist" containerID="f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.678999 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a"} err="failed to get container status \"f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a\": rpc error: code = NotFound desc = could not find container \"f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a\": container with ID starting with f68b0bacb0c936e649298dfc5afe1331b0f03f47e0d38d804d2a960522d89c2a not found: ID does not exist" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.696476 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb\") pod \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.696525 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc\") pod \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.696545 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config\") pod \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.696681 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk5jr\" (UniqueName: \"kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr\") pod \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\" (UID: \"1b467eaf-55a9-48a2-a852-008b6ebca0aa\") " Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.702708 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr" (OuterVolumeSpecName: "kube-api-access-lk5jr") pod "1b467eaf-55a9-48a2-a852-008b6ebca0aa" (UID: "1b467eaf-55a9-48a2-a852-008b6ebca0aa"). InnerVolumeSpecName "kube-api-access-lk5jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.745027 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b467eaf-55a9-48a2-a852-008b6ebca0aa" (UID: "1b467eaf-55a9-48a2-a852-008b6ebca0aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.751059 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config" (OuterVolumeSpecName: "config") pod "1b467eaf-55a9-48a2-a852-008b6ebca0aa" (UID: "1b467eaf-55a9-48a2-a852-008b6ebca0aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.791212 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b467eaf-55a9-48a2-a852-008b6ebca0aa" (UID: "1b467eaf-55a9-48a2-a852-008b6ebca0aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.798383 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk5jr\" (UniqueName: \"kubernetes.io/projected/1b467eaf-55a9-48a2-a852-008b6ebca0aa-kube-api-access-lk5jr\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.798422 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.798432 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.798441 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b467eaf-55a9-48a2-a852-008b6ebca0aa-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.949141 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:35 crc kubenswrapper[5018]: I1014 07:07:35.959294 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c5bcf59c-5nv7x"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.127370 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:07:36 crc kubenswrapper[5018]: W1014 07:07:36.128561 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57f1b1b7_2e2d_49a9_822f_8ffbd55251ad.slice/crio-2902f42e015becb2b9213efcf9bcf9bff6186fcadda218b81eba4aa14f10a327 WatchSource:0}: Error finding container 2902f42e015becb2b9213efcf9bcf9bff6186fcadda218b81eba4aa14f10a327: Status 404 returned error can't find the container with id 2902f42e015becb2b9213efcf9bcf9bff6186fcadda218b81eba4aa14f10a327 Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.304968 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.305343 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="dnsmasq-dns" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.305356 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="dnsmasq-dns" Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.305371 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="init" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.305377 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="init" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.307317 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" containerName="dnsmasq-dns" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.349074 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.349199 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.353777 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.353788 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-hkcfx" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.353900 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.353947 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.428362 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.428409 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wn8r\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.428435 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.428489 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.428607 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530264 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530342 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530386 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530434 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530453 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wn8r\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.530568 5018 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.530595 5018 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.530682 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift podName:50b722d0-93f3-41c5-a4e8-a314231f282b nodeName:}" failed. No retries permitted until 2025-10-14 07:07:37.030650108 +0000 UTC m=+1073.614696805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift") pod "swift-storage-0" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b") : configmap "swift-ring-files" not found Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530779 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530873 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.530875 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.548594 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wn8r\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.551544 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.615520 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b467eaf-55a9-48a2-a852-008b6ebca0aa" path="/var/lib/kubelet/pods/1b467eaf-55a9-48a2-a852-008b6ebca0aa/volumes" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.618434 5018 generic.go:334] "Generic (PLEG): container finished" podID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerID="cd4b9f95ca796b2e388b05c10f01ec12a79b441d49ce881004a2d675acfb271d" exitCode=0 Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.618512 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" event={"ID":"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad","Type":"ContainerDied","Data":"cd4b9f95ca796b2e388b05c10f01ec12a79b441d49ce881004a2d675acfb271d"} Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.618540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" event={"ID":"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad","Type":"ContainerStarted","Data":"2902f42e015becb2b9213efcf9bcf9bff6186fcadda218b81eba4aa14f10a327"} Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.812415 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6hsf6"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.814278 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.816660 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.816748 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.819066 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.842773 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6hsf6"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.850312 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-w4h5j"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.852547 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.857353 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-w4h5j"] Oct 14 07:07:36 crc kubenswrapper[5018]: E1014 07:07:36.894439 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ckhsh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ckhsh ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-6hsf6" podUID="32a4f96b-294c-4eb6-9fa7-749c95e19f1e" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.915557 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6hsf6"] Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940198 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940256 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940299 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940327 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hxth\" (UniqueName: \"kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940351 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940484 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940558 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940608 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940649 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940683 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940756 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940785 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckhsh\" (UniqueName: \"kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:36 crc kubenswrapper[5018]: I1014 07:07:36.940848 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.042742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hxth\" (UniqueName: \"kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043104 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043271 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043464 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043614 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043805 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043961 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.044002 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.043725 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.044293 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.044877 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.045122 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.045745 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckhsh\" (UniqueName: \"kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.046686 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.046855 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.046998 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047211 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047391 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047551 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047088 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047811 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.047953 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: E1014 07:07:37.048199 5018 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:07:37 crc kubenswrapper[5018]: E1014 07:07:37.048334 5018 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:07:37 crc kubenswrapper[5018]: E1014 07:07:37.048506 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift podName:50b722d0-93f3-41c5-a4e8-a314231f282b nodeName:}" failed. No retries permitted until 2025-10-14 07:07:38.048479959 +0000 UTC m=+1074.632526626 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift") pod "swift-storage-0" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b") : configmap "swift-ring-files" not found Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.048795 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.048816 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.049538 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.052012 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.062283 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hxth\" (UniqueName: \"kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth\") pod \"swift-ring-rebalance-w4h5j\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.069398 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckhsh\" (UniqueName: \"kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh\") pod \"swift-ring-rebalance-6hsf6\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.200425 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.420608 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.631434 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.632542 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" event={"ID":"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad","Type":"ContainerStarted","Data":"f94a6340ff419e5609d0db3cdb8fcc05fdb13312af60401fe1ada92fd6712189"} Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.632818 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.643286 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.654257 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" podStartSLOduration=2.654241109 podStartE2EDuration="2.654241109s" podCreationTimestamp="2025-10-14 07:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:07:37.652758628 +0000 UTC m=+1074.236805255" watchObservedRunningTime="2025-10-14 07:07:37.654241109 +0000 UTC m=+1074.238287736" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.688266 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-w4h5j"] Oct 14 07:07:37 crc kubenswrapper[5018]: W1014 07:07:37.688776 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7841210_eecb_475a_acf3_c1de58aa08a6.slice/crio-edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4 WatchSource:0}: Error finding container edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4: Status 404 returned error can't find the container with id edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4 Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763491 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763602 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763753 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckhsh\" (UniqueName: \"kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763803 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763916 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.763959 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.764045 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf\") pod \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\" (UID: \"32a4f96b-294c-4eb6-9fa7-749c95e19f1e\") " Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.765685 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.765830 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.766504 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts" (OuterVolumeSpecName: "scripts") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.770346 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.770680 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.771745 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh" (OuterVolumeSpecName: "kube-api-access-ckhsh") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "kube-api-access-ckhsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.772337 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "32a4f96b-294c-4eb6-9fa7-749c95e19f1e" (UID: "32a4f96b-294c-4eb6-9fa7-749c95e19f1e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.868914 5018 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.868991 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.869019 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckhsh\" (UniqueName: \"kubernetes.io/projected/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-kube-api-access-ckhsh\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.869071 5018 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.869096 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.869122 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:37 crc kubenswrapper[5018]: I1014 07:07:37.869148 5018 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/32a4f96b-294c-4eb6-9fa7-749c95e19f1e-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:38 crc kubenswrapper[5018]: I1014 07:07:38.076511 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:38 crc kubenswrapper[5018]: E1014 07:07:38.076686 5018 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:07:38 crc kubenswrapper[5018]: E1014 07:07:38.076699 5018 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:07:38 crc kubenswrapper[5018]: E1014 07:07:38.076746 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift podName:50b722d0-93f3-41c5-a4e8-a314231f282b nodeName:}" failed. No retries permitted until 2025-10-14 07:07:40.076732387 +0000 UTC m=+1076.660779014 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift") pod "swift-storage-0" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b") : configmap "swift-ring-files" not found Oct 14 07:07:38 crc kubenswrapper[5018]: I1014 07:07:38.641411 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-w4h5j" event={"ID":"d7841210-eecb-475a-acf3-c1de58aa08a6","Type":"ContainerStarted","Data":"edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4"} Oct 14 07:07:38 crc kubenswrapper[5018]: I1014 07:07:38.641438 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6hsf6" Oct 14 07:07:38 crc kubenswrapper[5018]: I1014 07:07:38.696280 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6hsf6"] Oct 14 07:07:38 crc kubenswrapper[5018]: I1014 07:07:38.703403 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-6hsf6"] Oct 14 07:07:39 crc kubenswrapper[5018]: I1014 07:07:39.139929 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:39 crc kubenswrapper[5018]: I1014 07:07:39.207792 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 07:07:40 crc kubenswrapper[5018]: I1014 07:07:40.113106 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:40 crc kubenswrapper[5018]: E1014 07:07:40.113321 5018 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:07:40 crc kubenswrapper[5018]: E1014 07:07:40.113495 5018 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:07:40 crc kubenswrapper[5018]: E1014 07:07:40.113553 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift podName:50b722d0-93f3-41c5-a4e8-a314231f282b nodeName:}" failed. No retries permitted until 2025-10-14 07:07:44.113537161 +0000 UTC m=+1080.697583788 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift") pod "swift-storage-0" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b") : configmap "swift-ring-files" not found Oct 14 07:07:40 crc kubenswrapper[5018]: I1014 07:07:40.624171 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a4f96b-294c-4eb6-9fa7-749c95e19f1e" path="/var/lib/kubelet/pods/32a4f96b-294c-4eb6-9fa7-749c95e19f1e/volumes" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.669072 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25658566-e65b-48fe-aa02-1df2ab82b526","Type":"ContainerStarted","Data":"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1"} Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.670035 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.672646 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-w4h5j" event={"ID":"d7841210-eecb-475a-acf3-c1de58aa08a6","Type":"ContainerStarted","Data":"2857b484232f0cbf1617641f754477a7d6b6fa7212de8ab2bc4d14513764179d"} Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.694104 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.339113352 podStartE2EDuration="37.694080519s" podCreationTimestamp="2025-10-14 07:07:04 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.797335581 +0000 UTC m=+1051.381382208" lastFinishedPulling="2025-10-14 07:07:41.152302728 +0000 UTC m=+1077.736349375" observedRunningTime="2025-10-14 07:07:41.691300262 +0000 UTC m=+1078.275346899" watchObservedRunningTime="2025-10-14 07:07:41.694080519 +0000 UTC m=+1078.278127186" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.706848 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.707137 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.721128 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-w4h5j" podStartSLOduration=2.232662537 podStartE2EDuration="5.721066731s" podCreationTimestamp="2025-10-14 07:07:36 +0000 UTC" firstStartedPulling="2025-10-14 07:07:37.691062822 +0000 UTC m=+1074.275109449" lastFinishedPulling="2025-10-14 07:07:41.179467016 +0000 UTC m=+1077.763513643" observedRunningTime="2025-10-14 07:07:41.719078507 +0000 UTC m=+1078.303125174" watchObservedRunningTime="2025-10-14 07:07:41.721066731 +0000 UTC m=+1078.305113398" Oct 14 07:07:41 crc kubenswrapper[5018]: I1014 07:07:41.780036 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 07:07:42 crc kubenswrapper[5018]: I1014 07:07:42.762402 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.284113 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-s4jrd"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.285992 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.296699 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s4jrd"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.378536 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmv4c\" (UniqueName: \"kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c\") pod \"keystone-db-create-s4jrd\" (UID: \"1801677b-8aa6-4c0f-9f88-66bea7da4835\") " pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.479995 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmv4c\" (UniqueName: \"kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c\") pod \"keystone-db-create-s4jrd\" (UID: \"1801677b-8aa6-4c0f-9f88-66bea7da4835\") " pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.502574 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmv4c\" (UniqueName: \"kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c\") pod \"keystone-db-create-s4jrd\" (UID: \"1801677b-8aa6-4c0f-9f88-66bea7da4835\") " pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.559411 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c4hlc"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.560370 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.578911 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c4hlc"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.607089 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.684749 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfrds\" (UniqueName: \"kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds\") pod \"placement-db-create-c4hlc\" (UID: \"350454c7-242f-4dab-bb12-21abca9db8ce\") " pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.764041 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2km9n"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.765257 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2km9n" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.788742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfrds\" (UniqueName: \"kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds\") pod \"placement-db-create-c4hlc\" (UID: \"350454c7-242f-4dab-bb12-21abca9db8ce\") " pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.789813 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2km9n"] Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.821287 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfrds\" (UniqueName: \"kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds\") pod \"placement-db-create-c4hlc\" (UID: \"350454c7-242f-4dab-bb12-21abca9db8ce\") " pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.888633 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.890063 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkx5v\" (UniqueName: \"kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v\") pod \"glance-db-create-2km9n\" (UID: \"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18\") " pod="openstack/glance-db-create-2km9n" Oct 14 07:07:43 crc kubenswrapper[5018]: I1014 07:07:43.991862 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkx5v\" (UniqueName: \"kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v\") pod \"glance-db-create-2km9n\" (UID: \"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18\") " pod="openstack/glance-db-create-2km9n" Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.012478 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkx5v\" (UniqueName: \"kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v\") pod \"glance-db-create-2km9n\" (UID: \"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18\") " pod="openstack/glance-db-create-2km9n" Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.064511 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s4jrd"] Oct 14 07:07:44 crc kubenswrapper[5018]: W1014 07:07:44.084277 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1801677b_8aa6_4c0f_9f88_66bea7da4835.slice/crio-bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9 WatchSource:0}: Error finding container bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9: Status 404 returned error can't find the container with id bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9 Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.094980 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2km9n" Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.194003 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:44 crc kubenswrapper[5018]: E1014 07:07:44.194193 5018 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:07:44 crc kubenswrapper[5018]: E1014 07:07:44.194346 5018 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:07:44 crc kubenswrapper[5018]: E1014 07:07:44.194406 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift podName:50b722d0-93f3-41c5-a4e8-a314231f282b nodeName:}" failed. No retries permitted until 2025-10-14 07:07:52.19438848 +0000 UTC m=+1088.778435107 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift") pod "swift-storage-0" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b") : configmap "swift-ring-files" not found Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.358213 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c4hlc"] Oct 14 07:07:44 crc kubenswrapper[5018]: W1014 07:07:44.405580 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod350454c7_242f_4dab_bb12_21abca9db8ce.slice/crio-64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986 WatchSource:0}: Error finding container 64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986: Status 404 returned error can't find the container with id 64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986 Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.561424 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2km9n"] Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.697366 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2km9n" event={"ID":"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18","Type":"ContainerStarted","Data":"7e2deeebe20f4b84039e99b19f232ad2bae7be714d84e4ce38a206ec74ec859c"} Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.698924 5018 generic.go:334] "Generic (PLEG): container finished" podID="350454c7-242f-4dab-bb12-21abca9db8ce" containerID="553347b1b80d8f4b315497ba5161f42a2b187bace35cb9dba665dc45ea96194d" exitCode=0 Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.698969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c4hlc" event={"ID":"350454c7-242f-4dab-bb12-21abca9db8ce","Type":"ContainerDied","Data":"553347b1b80d8f4b315497ba5161f42a2b187bace35cb9dba665dc45ea96194d"} Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.698986 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c4hlc" event={"ID":"350454c7-242f-4dab-bb12-21abca9db8ce","Type":"ContainerStarted","Data":"64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986"} Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.709394 5018 generic.go:334] "Generic (PLEG): container finished" podID="1801677b-8aa6-4c0f-9f88-66bea7da4835" containerID="72aaf2f3ea790510bd282622e3cc4b3f450e19ea9c51089f9720e414ce4c5e79" exitCode=0 Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.709451 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4jrd" event={"ID":"1801677b-8aa6-4c0f-9f88-66bea7da4835","Type":"ContainerDied","Data":"72aaf2f3ea790510bd282622e3cc4b3f450e19ea9c51089f9720e414ce4c5e79"} Oct 14 07:07:44 crc kubenswrapper[5018]: I1014 07:07:44.709809 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4jrd" event={"ID":"1801677b-8aa6-4c0f-9f88-66bea7da4835","Type":"ContainerStarted","Data":"bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9"} Oct 14 07:07:45 crc kubenswrapper[5018]: I1014 07:07:45.619847 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:07:45 crc kubenswrapper[5018]: I1014 07:07:45.699424 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:45 crc kubenswrapper[5018]: I1014 07:07:45.702150 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="dnsmasq-dns" containerID="cri-o://946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3" gracePeriod=10 Oct 14 07:07:45 crc kubenswrapper[5018]: I1014 07:07:45.736390 5018 generic.go:334] "Generic (PLEG): container finished" podID="2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" containerID="f106d8478ac0f489c8ee548d9ecffc0f6885c004e9abab5c884eddda23eedfea" exitCode=0 Oct 14 07:07:45 crc kubenswrapper[5018]: I1014 07:07:45.737061 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2km9n" event={"ID":"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18","Type":"ContainerDied","Data":"f106d8478ac0f489c8ee548d9ecffc0f6885c004e9abab5c884eddda23eedfea"} Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.262927 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.267840 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.273435 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329224 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb\") pod \"e9dc8f93-29e8-482a-9a26-418541db4e61\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329290 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfrds\" (UniqueName: \"kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds\") pod \"350454c7-242f-4dab-bb12-21abca9db8ce\" (UID: \"350454c7-242f-4dab-bb12-21abca9db8ce\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329360 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrzzp\" (UniqueName: \"kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp\") pod \"e9dc8f93-29e8-482a-9a26-418541db4e61\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329455 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmv4c\" (UniqueName: \"kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c\") pod \"1801677b-8aa6-4c0f-9f88-66bea7da4835\" (UID: \"1801677b-8aa6-4c0f-9f88-66bea7da4835\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329481 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb\") pod \"e9dc8f93-29e8-482a-9a26-418541db4e61\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329514 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc\") pod \"e9dc8f93-29e8-482a-9a26-418541db4e61\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.329560 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config\") pod \"e9dc8f93-29e8-482a-9a26-418541db4e61\" (UID: \"e9dc8f93-29e8-482a-9a26-418541db4e61\") " Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.341786 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c" (OuterVolumeSpecName: "kube-api-access-bmv4c") pod "1801677b-8aa6-4c0f-9f88-66bea7da4835" (UID: "1801677b-8aa6-4c0f-9f88-66bea7da4835"). InnerVolumeSpecName "kube-api-access-bmv4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.341829 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds" (OuterVolumeSpecName: "kube-api-access-qfrds") pod "350454c7-242f-4dab-bb12-21abca9db8ce" (UID: "350454c7-242f-4dab-bb12-21abca9db8ce"). InnerVolumeSpecName "kube-api-access-qfrds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.352906 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp" (OuterVolumeSpecName: "kube-api-access-mrzzp") pod "e9dc8f93-29e8-482a-9a26-418541db4e61" (UID: "e9dc8f93-29e8-482a-9a26-418541db4e61"). InnerVolumeSpecName "kube-api-access-mrzzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.370426 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e9dc8f93-29e8-482a-9a26-418541db4e61" (UID: "e9dc8f93-29e8-482a-9a26-418541db4e61"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.383046 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9dc8f93-29e8-482a-9a26-418541db4e61" (UID: "e9dc8f93-29e8-482a-9a26-418541db4e61"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.383727 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9dc8f93-29e8-482a-9a26-418541db4e61" (UID: "e9dc8f93-29e8-482a-9a26-418541db4e61"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.389521 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config" (OuterVolumeSpecName: "config") pod "e9dc8f93-29e8-482a-9a26-418541db4e61" (UID: "e9dc8f93-29e8-482a-9a26-418541db4e61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430898 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430931 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfrds\" (UniqueName: \"kubernetes.io/projected/350454c7-242f-4dab-bb12-21abca9db8ce-kube-api-access-qfrds\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430945 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrzzp\" (UniqueName: \"kubernetes.io/projected/e9dc8f93-29e8-482a-9a26-418541db4e61-kube-api-access-mrzzp\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430956 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmv4c\" (UniqueName: \"kubernetes.io/projected/1801677b-8aa6-4c0f-9f88-66bea7da4835-kube-api-access-bmv4c\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430966 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430976 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.430986 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9dc8f93-29e8-482a-9a26-418541db4e61-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.748919 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c4hlc" event={"ID":"350454c7-242f-4dab-bb12-21abca9db8ce","Type":"ContainerDied","Data":"64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986"} Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.748979 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64ece06be149106f61d2eae7522b9e2685f48fb1de69b1e64e004c3d576a1986" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.749071 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c4hlc" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.751572 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s4jrd" event={"ID":"1801677b-8aa6-4c0f-9f88-66bea7da4835","Type":"ContainerDied","Data":"bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9"} Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.751654 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdaa1508d27fbd1d9c951f5cb600cf2cde61a6bdf94f01eeb404e7e6334f6fc9" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.751658 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s4jrd" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.756482 5018 generic.go:334] "Generic (PLEG): container finished" podID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerID="946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3" exitCode=0 Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.756537 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.756548 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" event={"ID":"e9dc8f93-29e8-482a-9a26-418541db4e61","Type":"ContainerDied","Data":"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3"} Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.756595 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d49c4d845-m7xc5" event={"ID":"e9dc8f93-29e8-482a-9a26-418541db4e61","Type":"ContainerDied","Data":"ec522f3017902d9e0ede06a300a45929d0fe3c6b65b874a2f053f52ad45f61f4"} Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.756632 5018 scope.go:117] "RemoveContainer" containerID="946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.784773 5018 scope.go:117] "RemoveContainer" containerID="8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.786859 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.796743 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d49c4d845-m7xc5"] Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.807914 5018 scope.go:117] "RemoveContainer" containerID="946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3" Oct 14 07:07:46 crc kubenswrapper[5018]: E1014 07:07:46.809333 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3\": container with ID starting with 946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3 not found: ID does not exist" containerID="946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.809372 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3"} err="failed to get container status \"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3\": rpc error: code = NotFound desc = could not find container \"946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3\": container with ID starting with 946926a180b5f8c93a958b809d37fb65c878e424e0f89e795903e6d4ad55c1d3 not found: ID does not exist" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.809399 5018 scope.go:117] "RemoveContainer" containerID="8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767" Oct 14 07:07:46 crc kubenswrapper[5018]: E1014 07:07:46.809713 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767\": container with ID starting with 8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767 not found: ID does not exist" containerID="8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767" Oct 14 07:07:46 crc kubenswrapper[5018]: I1014 07:07:46.809751 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767"} err="failed to get container status \"8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767\": rpc error: code = NotFound desc = could not find container \"8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767\": container with ID starting with 8a6879f617260caf526ec0eb20e94c379076016e590228235dc5425ac25cc767 not found: ID does not exist" Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.066886 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2km9n" Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.142042 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkx5v\" (UniqueName: \"kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v\") pod \"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18\" (UID: \"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18\") " Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.146818 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v" (OuterVolumeSpecName: "kube-api-access-nkx5v") pod "2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" (UID: "2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18"). InnerVolumeSpecName "kube-api-access-nkx5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.243819 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkx5v\" (UniqueName: \"kubernetes.io/projected/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18-kube-api-access-nkx5v\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.767704 5018 generic.go:334] "Generic (PLEG): container finished" podID="d7841210-eecb-475a-acf3-c1de58aa08a6" containerID="2857b484232f0cbf1617641f754477a7d6b6fa7212de8ab2bc4d14513764179d" exitCode=0 Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.767811 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-w4h5j" event={"ID":"d7841210-eecb-475a-acf3-c1de58aa08a6","Type":"ContainerDied","Data":"2857b484232f0cbf1617641f754477a7d6b6fa7212de8ab2bc4d14513764179d"} Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.769560 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2km9n" event={"ID":"2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18","Type":"ContainerDied","Data":"7e2deeebe20f4b84039e99b19f232ad2bae7be714d84e4ce38a206ec74ec859c"} Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.769610 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e2deeebe20f4b84039e99b19f232ad2bae7be714d84e4ce38a206ec74ec859c" Oct 14 07:07:47 crc kubenswrapper[5018]: I1014 07:07:47.769684 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2km9n" Oct 14 07:07:48 crc kubenswrapper[5018]: I1014 07:07:48.375843 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 07:07:48 crc kubenswrapper[5018]: I1014 07:07:48.616765 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" path="/var/lib/kubelet/pods/e9dc8f93-29e8-482a-9a26-418541db4e61/volumes" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.183981 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.277813 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hxth\" (UniqueName: \"kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.277864 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.277951 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.277976 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.278025 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.278055 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.278102 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf\") pod \"d7841210-eecb-475a-acf3-c1de58aa08a6\" (UID: \"d7841210-eecb-475a-acf3-c1de58aa08a6\") " Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.278773 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.279552 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.285724 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth" (OuterVolumeSpecName: "kube-api-access-5hxth") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "kube-api-access-5hxth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.288333 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.303562 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts" (OuterVolumeSpecName: "scripts") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.303722 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.306683 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d7841210-eecb-475a-acf3-c1de58aa08a6" (UID: "d7841210-eecb-475a-acf3-c1de58aa08a6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380291 5018 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380329 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380341 5018 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380353 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hxth\" (UniqueName: \"kubernetes.io/projected/d7841210-eecb-475a-acf3-c1de58aa08a6-kube-api-access-5hxth\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380367 5018 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d7841210-eecb-475a-acf3-c1de58aa08a6-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380378 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d7841210-eecb-475a-acf3-c1de58aa08a6-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.380389 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7841210-eecb-475a-acf3-c1de58aa08a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.796557 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-w4h5j" event={"ID":"d7841210-eecb-475a-acf3-c1de58aa08a6","Type":"ContainerDied","Data":"edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4"} Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.796683 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edf33e0b424c1053323fe68b003800804217ac71563ce943a8ac75bc0ea706e4" Oct 14 07:07:49 crc kubenswrapper[5018]: I1014 07:07:49.796699 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w4h5j" Oct 14 07:07:52 crc kubenswrapper[5018]: I1014 07:07:52.236486 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:52 crc kubenswrapper[5018]: I1014 07:07:52.248304 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"swift-storage-0\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " pod="openstack/swift-storage-0" Oct 14 07:07:52 crc kubenswrapper[5018]: I1014 07:07:52.271958 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:07:52 crc kubenswrapper[5018]: I1014 07:07:52.909584 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296468 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-62dd-account-create-mk5wq"] Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296763 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1801677b-8aa6-4c0f-9f88-66bea7da4835" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296774 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1801677b-8aa6-4c0f-9f88-66bea7da4835" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296795 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="init" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296800 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="init" Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296810 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7841210-eecb-475a-acf3-c1de58aa08a6" containerName="swift-ring-rebalance" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296816 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7841210-eecb-475a-acf3-c1de58aa08a6" containerName="swift-ring-rebalance" Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296832 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350454c7-242f-4dab-bb12-21abca9db8ce" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296838 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="350454c7-242f-4dab-bb12-21abca9db8ce" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296849 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296855 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: E1014 07:07:53.296866 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="dnsmasq-dns" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.296871 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="dnsmasq-dns" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297054 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1801677b-8aa6-4c0f-9f88-66bea7da4835" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297064 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="350454c7-242f-4dab-bb12-21abca9db8ce" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297078 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7841210-eecb-475a-acf3-c1de58aa08a6" containerName="swift-ring-rebalance" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297087 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" containerName="mariadb-database-create" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297098 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9dc8f93-29e8-482a-9a26-418541db4e61" containerName="dnsmasq-dns" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.297595 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.299574 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.316463 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-62dd-account-create-mk5wq"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.357340 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4jq2\" (UniqueName: \"kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2\") pod \"keystone-62dd-account-create-mk5wq\" (UID: \"57e80324-584c-4cfd-b71e-11adfe900797\") " pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.458764 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4jq2\" (UniqueName: \"kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2\") pod \"keystone-62dd-account-create-mk5wq\" (UID: \"57e80324-584c-4cfd-b71e-11adfe900797\") " pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.503703 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4jq2\" (UniqueName: \"kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2\") pod \"keystone-62dd-account-create-mk5wq\" (UID: \"57e80324-584c-4cfd-b71e-11adfe900797\") " pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.602072 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d72f-account-create-897bn"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.603236 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.606117 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.617212 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d72f-account-create-897bn"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.637756 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.663127 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nl4c\" (UniqueName: \"kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c\") pod \"placement-d72f-account-create-897bn\" (UID: \"09ac7a1f-c474-4f6e-9a09-e545d75e33c8\") " pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.764871 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nl4c\" (UniqueName: \"kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c\") pod \"placement-d72f-account-create-897bn\" (UID: \"09ac7a1f-c474-4f6e-9a09-e545d75e33c8\") " pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.804720 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-dba2-account-create-8x996"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.805572 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nl4c\" (UniqueName: \"kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c\") pod \"placement-d72f-account-create-897bn\" (UID: \"09ac7a1f-c474-4f6e-9a09-e545d75e33c8\") " pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.805691 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.810047 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.811225 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dba2-account-create-8x996"] Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.831562 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"9317f296aa82a0881db11bdcabbe4cf06b4ce838519fad7b7e76b7ea26d99ee0"} Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.866412 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x89zf\" (UniqueName: \"kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf\") pod \"glance-dba2-account-create-8x996\" (UID: \"55477fe1-67aa-45f6-ba53-7a39c81acfd1\") " pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.918962 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.968151 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x89zf\" (UniqueName: \"kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf\") pod \"glance-dba2-account-create-8x996\" (UID: \"55477fe1-67aa-45f6-ba53-7a39c81acfd1\") " pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:53 crc kubenswrapper[5018]: I1014 07:07:53.994552 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x89zf\" (UniqueName: \"kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf\") pod \"glance-dba2-account-create-8x996\" (UID: \"55477fe1-67aa-45f6-ba53-7a39c81acfd1\") " pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.056873 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-62dd-account-create-mk5wq"] Oct 14 07:07:54 crc kubenswrapper[5018]: W1014 07:07:54.070904 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57e80324_584c_4cfd_b71e_11adfe900797.slice/crio-7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91 WatchSource:0}: Error finding container 7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91: Status 404 returned error can't find the container with id 7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91 Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.131362 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.370514 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d72f-account-create-897bn"] Oct 14 07:07:54 crc kubenswrapper[5018]: W1014 07:07:54.380081 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09ac7a1f_c474_4f6e_9a09_e545d75e33c8.slice/crio-809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5 WatchSource:0}: Error finding container 809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5: Status 404 returned error can't find the container with id 809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5 Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.713288 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-dba2-account-create-8x996"] Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.841702 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dba2-account-create-8x996" event={"ID":"55477fe1-67aa-45f6-ba53-7a39c81acfd1","Type":"ContainerStarted","Data":"17b43a5630f08e19b3f29b5d53830f56af0ee5798e578841a3d414292d91300e"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.847710 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d72f-account-create-897bn" event={"ID":"09ac7a1f-c474-4f6e-9a09-e545d75e33c8","Type":"ContainerStarted","Data":"1889b1037262e6670ececf5f3cc0fad455f0f676c091666af0087aa6c4a7318a"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.847843 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d72f-account-create-897bn" event={"ID":"09ac7a1f-c474-4f6e-9a09-e545d75e33c8","Type":"ContainerStarted","Data":"809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.853403 5018 generic.go:334] "Generic (PLEG): container finished" podID="57e80324-584c-4cfd-b71e-11adfe900797" containerID="0090c0a99892af1c98856d4013a796c1ad868a038ad76ec6b002625b0cb74e54" exitCode=0 Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.853562 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-62dd-account-create-mk5wq" event={"ID":"57e80324-584c-4cfd-b71e-11adfe900797","Type":"ContainerDied","Data":"0090c0a99892af1c98856d4013a796c1ad868a038ad76ec6b002625b0cb74e54"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.853886 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-62dd-account-create-mk5wq" event={"ID":"57e80324-584c-4cfd-b71e-11adfe900797","Type":"ContainerStarted","Data":"7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.856684 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"422e0f3e0acbeb85eabbf4a53967dd8d157d1f9364ac749656bd718ee3f5b1e7"} Oct 14 07:07:54 crc kubenswrapper[5018]: I1014 07:07:54.866451 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d72f-account-create-897bn" podStartSLOduration=1.8664277519999999 podStartE2EDuration="1.866427752s" podCreationTimestamp="2025-10-14 07:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:07:54.864261473 +0000 UTC m=+1091.448308110" watchObservedRunningTime="2025-10-14 07:07:54.866427752 +0000 UTC m=+1091.450474419" Oct 14 07:07:55 crc kubenswrapper[5018]: I1014 07:07:55.103421 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 07:07:55 crc kubenswrapper[5018]: I1014 07:07:55.873075 5018 generic.go:334] "Generic (PLEG): container finished" podID="09ac7a1f-c474-4f6e-9a09-e545d75e33c8" containerID="1889b1037262e6670ececf5f3cc0fad455f0f676c091666af0087aa6c4a7318a" exitCode=0 Oct 14 07:07:55 crc kubenswrapper[5018]: I1014 07:07:55.873223 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d72f-account-create-897bn" event={"ID":"09ac7a1f-c474-4f6e-9a09-e545d75e33c8","Type":"ContainerDied","Data":"1889b1037262e6670ececf5f3cc0fad455f0f676c091666af0087aa6c4a7318a"} Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.208863 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.312941 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4jq2\" (UniqueName: \"kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2\") pod \"57e80324-584c-4cfd-b71e-11adfe900797\" (UID: \"57e80324-584c-4cfd-b71e-11adfe900797\") " Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.324863 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2" (OuterVolumeSpecName: "kube-api-access-x4jq2") pod "57e80324-584c-4cfd-b71e-11adfe900797" (UID: "57e80324-584c-4cfd-b71e-11adfe900797"). InnerVolumeSpecName "kube-api-access-x4jq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.416608 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4jq2\" (UniqueName: \"kubernetes.io/projected/57e80324-584c-4cfd-b71e-11adfe900797-kube-api-access-x4jq2\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.890973 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-62dd-account-create-mk5wq" event={"ID":"57e80324-584c-4cfd-b71e-11adfe900797","Type":"ContainerDied","Data":"7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91"} Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.891579 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7461e420c3d723162cab1a710607bde97f6a3f7f69368fd600b05f00d8abea91" Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.891063 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-62dd-account-create-mk5wq" Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.902187 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"ee71c8af72a00afcf6ab9a1af606dec3a2657101b03f635f39d3c3a416113250"} Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.902419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"a9db9d72a3a3664bb31edfeb916708e025a2ea1bc87363e6693b7d4ae2caf69c"} Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.902576 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"b83fc34a9d177c956e778ff75097967bb0fc448f9d4a508069b6032e5cd85729"} Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.905546 5018 generic.go:334] "Generic (PLEG): container finished" podID="55477fe1-67aa-45f6-ba53-7a39c81acfd1" containerID="295bcb9c9f247ccad615ca5131bff760f8822e3791874a3fa3874a7d346939e7" exitCode=0 Oct 14 07:07:56 crc kubenswrapper[5018]: I1014 07:07:56.905840 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dba2-account-create-8x996" event={"ID":"55477fe1-67aa-45f6-ba53-7a39c81acfd1","Type":"ContainerDied","Data":"295bcb9c9f247ccad615ca5131bff760f8822e3791874a3fa3874a7d346939e7"} Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.290393 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.439331 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nl4c\" (UniqueName: \"kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c\") pod \"09ac7a1f-c474-4f6e-9a09-e545d75e33c8\" (UID: \"09ac7a1f-c474-4f6e-9a09-e545d75e33c8\") " Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.444926 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c" (OuterVolumeSpecName: "kube-api-access-4nl4c") pod "09ac7a1f-c474-4f6e-9a09-e545d75e33c8" (UID: "09ac7a1f-c474-4f6e-9a09-e545d75e33c8"). InnerVolumeSpecName "kube-api-access-4nl4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.541415 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nl4c\" (UniqueName: \"kubernetes.io/projected/09ac7a1f-c474-4f6e-9a09-e545d75e33c8-kube-api-access-4nl4c\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.917928 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d72f-account-create-897bn" event={"ID":"09ac7a1f-c474-4f6e-9a09-e545d75e33c8","Type":"ContainerDied","Data":"809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5"} Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.917980 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="809b2189fbc806f9a48d70401ff19b1851f85abd7581de42668dd8b87af358f5" Oct 14 07:07:57 crc kubenswrapper[5018]: I1014 07:07:57.917989 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d72f-account-create-897bn" Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.243182 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.255372 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x89zf\" (UniqueName: \"kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf\") pod \"55477fe1-67aa-45f6-ba53-7a39c81acfd1\" (UID: \"55477fe1-67aa-45f6-ba53-7a39c81acfd1\") " Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.278116 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf" (OuterVolumeSpecName: "kube-api-access-x89zf") pod "55477fe1-67aa-45f6-ba53-7a39c81acfd1" (UID: "55477fe1-67aa-45f6-ba53-7a39c81acfd1"). InnerVolumeSpecName "kube-api-access-x89zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.356936 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x89zf\" (UniqueName: \"kubernetes.io/projected/55477fe1-67aa-45f6-ba53-7a39c81acfd1-kube-api-access-x89zf\") on node \"crc\" DevicePath \"\"" Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.930070 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"b05e853fcda282db718ede400b9daeb449fd5e7cd0b75409d593e3b167779576"} Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.930117 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"ebd9c6ab8467a65fedbce41b1ce78986ad397f0ca50597aaea953816e194054d"} Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.930133 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"80c27ba55cfa5a682b1e2072de8e447382019d0b79b38cf3f4de1b47c35a429d"} Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.930148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"e58847c0dec122f46ec351ebf113ea26a71f6dd0002b59bf78203f30ea173dca"} Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.932225 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-dba2-account-create-8x996" event={"ID":"55477fe1-67aa-45f6-ba53-7a39c81acfd1","Type":"ContainerDied","Data":"17b43a5630f08e19b3f29b5d53830f56af0ee5798e578841a3d414292d91300e"} Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.932254 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-dba2-account-create-8x996" Oct 14 07:07:58 crc kubenswrapper[5018]: I1014 07:07:58.932271 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17b43a5630f08e19b3f29b5d53830f56af0ee5798e578841a3d414292d91300e" Oct 14 07:07:59 crc kubenswrapper[5018]: I1014 07:07:59.943865 5018 generic.go:334] "Generic (PLEG): container finished" podID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerID="124f8aee2f1028a966a6895f7aa8cc6cea30f99e63669865c80d68ed595f9d7f" exitCode=0 Oct 14 07:07:59 crc kubenswrapper[5018]: I1014 07:07:59.943947 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerDied","Data":"124f8aee2f1028a966a6895f7aa8cc6cea30f99e63669865c80d68ed595f9d7f"} Oct 14 07:07:59 crc kubenswrapper[5018]: I1014 07:07:59.946140 5018 generic.go:334] "Generic (PLEG): container finished" podID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerID="6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a" exitCode=0 Oct 14 07:07:59 crc kubenswrapper[5018]: I1014 07:07:59.946217 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerDied","Data":"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a"} Oct 14 07:08:00 crc kubenswrapper[5018]: I1014 07:08:00.051274 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gs786" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" probeResult="failure" output=< Oct 14 07:08:00 crc kubenswrapper[5018]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 07:08:00 crc kubenswrapper[5018]: > Oct 14 07:08:00 crc kubenswrapper[5018]: I1014 07:08:00.964263 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"1c34c44d438a3d6af5f9816098723851b45124c4a7a2d7e1641ec047b86764eb"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.978000 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"7464ef303afd6563d9b11c50e58be39b3a6ce6d6bc0c340a249ad132b5dfdcba"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.978538 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"969d5f8aadb42ee0593e5d86418c4fdb29ca61ce0f36d2a945eec3737e42b189"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.978557 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"ec680e0a29290410e29c3d3731406e8216b29a7138c4bf28271899256ee732cc"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.978568 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"599bb15e72a850b4e7c1ebc86b49f02ba6953b1cd2083f0f540b8f3b91ff08c0"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.978578 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"963b3d72e03491fd9f8e960e77328e431e858c28e5183636d20edce5d6da90ec"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.980148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerStarted","Data":"db5a58f0d517ed36e8022372c1d0cbb0ac63dbc1e9c0b3d9627b4e9590b35c60"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.980318 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.982852 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerStarted","Data":"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804"} Oct 14 07:08:01 crc kubenswrapper[5018]: I1014 07:08:01.983055 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 07:08:02 crc kubenswrapper[5018]: I1014 07:08:02.006066 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.080679845 podStartE2EDuration="1m4.006035149s" podCreationTimestamp="2025-10-14 07:06:58 +0000 UTC" firstStartedPulling="2025-10-14 07:07:14.497936741 +0000 UTC m=+1051.081983368" lastFinishedPulling="2025-10-14 07:07:25.423292045 +0000 UTC m=+1062.007338672" observedRunningTime="2025-10-14 07:08:02.001033112 +0000 UTC m=+1098.585079749" watchObservedRunningTime="2025-10-14 07:08:02.006035149 +0000 UTC m=+1098.590081776" Oct 14 07:08:02 crc kubenswrapper[5018]: I1014 07:08:02.025945 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.490502133 podStartE2EDuration="1m4.025925607s" podCreationTimestamp="2025-10-14 07:06:58 +0000 UTC" firstStartedPulling="2025-10-14 07:07:11.634909479 +0000 UTC m=+1048.218956106" lastFinishedPulling="2025-10-14 07:07:25.170332953 +0000 UTC m=+1061.754379580" observedRunningTime="2025-10-14 07:08:02.025393062 +0000 UTC m=+1098.609439689" watchObservedRunningTime="2025-10-14 07:08:02.025925607 +0000 UTC m=+1098.609972244" Oct 14 07:08:02 crc kubenswrapper[5018]: I1014 07:08:02.463705 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:08:02 crc kubenswrapper[5018]: I1014 07:08:02.463789 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:08:02 crc kubenswrapper[5018]: I1014 07:08:02.999137 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerStarted","Data":"f75caf8dbbd7c4743e32fa90f2af02023f5f583c5e27a58a4d2ae6216d5c09b9"} Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.054116 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.113966415 podStartE2EDuration="28.054086642s" podCreationTimestamp="2025-10-14 07:07:35 +0000 UTC" firstStartedPulling="2025-10-14 07:07:52.903923022 +0000 UTC m=+1089.487969689" lastFinishedPulling="2025-10-14 07:07:59.844043289 +0000 UTC m=+1096.428089916" observedRunningTime="2025-10-14 07:08:03.046465112 +0000 UTC m=+1099.630511759" watchObservedRunningTime="2025-10-14 07:08:03.054086642 +0000 UTC m=+1099.638133279" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.347141 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:03 crc kubenswrapper[5018]: E1014 07:08:03.347780 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55477fe1-67aa-45f6-ba53-7a39c81acfd1" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.347799 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="55477fe1-67aa-45f6-ba53-7a39c81acfd1" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: E1014 07:08:03.347817 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ac7a1f-c474-4f6e-9a09-e545d75e33c8" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.347826 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ac7a1f-c474-4f6e-9a09-e545d75e33c8" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: E1014 07:08:03.347841 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e80324-584c-4cfd-b71e-11adfe900797" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.347851 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e80324-584c-4cfd-b71e-11adfe900797" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.348055 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ac7a1f-c474-4f6e-9a09-e545d75e33c8" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.348084 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e80324-584c-4cfd-b71e-11adfe900797" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.348097 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="55477fe1-67aa-45f6-ba53-7a39c81acfd1" containerName="mariadb-account-create" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.349139 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.353512 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.363333 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448437 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448519 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nppb5\" (UniqueName: \"kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448656 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448713 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.448739 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550454 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550524 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nppb5\" (UniqueName: \"kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550548 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550595 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550643 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.550665 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.551541 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.551584 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.551590 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.551713 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.552073 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.568901 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nppb5\" (UniqueName: \"kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5\") pod \"dnsmasq-dns-d9ddcb47c-qm2ls\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.679704 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.941068 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-77v4n"] Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.942419 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.944539 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.944755 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pj22j" Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.952030 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-77v4n"] Oct 14 07:08:03 crc kubenswrapper[5018]: I1014 07:08:03.985443 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:03 crc kubenswrapper[5018]: W1014 07:08:03.985949 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b17231c_1edb_4639_8333_678e3434f4dd.slice/crio-bf7603f5e713fba4b1f9dd355a1af92e7fd57dba43286967a1fe9629fc4abfd6 WatchSource:0}: Error finding container bf7603f5e713fba4b1f9dd355a1af92e7fd57dba43286967a1fe9629fc4abfd6: Status 404 returned error can't find the container with id bf7603f5e713fba4b1f9dd355a1af92e7fd57dba43286967a1fe9629fc4abfd6 Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.012762 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" event={"ID":"5b17231c-1edb-4639-8333-678e3434f4dd","Type":"ContainerStarted","Data":"bf7603f5e713fba4b1f9dd355a1af92e7fd57dba43286967a1fe9629fc4abfd6"} Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.063294 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.063358 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.063400 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxkxd\" (UniqueName: \"kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.063501 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.164736 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.165142 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.165179 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.165222 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxkxd\" (UniqueName: \"kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.170163 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.170744 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.171859 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.183046 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxkxd\" (UniqueName: \"kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd\") pod \"glance-db-sync-77v4n\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.263574 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:04 crc kubenswrapper[5018]: I1014 07:08:04.769205 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-77v4n"] Oct 14 07:08:04 crc kubenswrapper[5018]: W1014 07:08:04.776697 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaad96861_15f8_4715_8b7c_927ae48fe59c.slice/crio-2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a WatchSource:0}: Error finding container 2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a: Status 404 returned error can't find the container with id 2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.063579 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gs786" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" probeResult="failure" output=< Oct 14 07:08:05 crc kubenswrapper[5018]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 07:08:05 crc kubenswrapper[5018]: > Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.066867 5018 generic.go:334] "Generic (PLEG): container finished" podID="5b17231c-1edb-4639-8333-678e3434f4dd" containerID="60a1142fe263841ea85847e103545520b46aee8733cb982cb5e8bd7b8fc70379" exitCode=0 Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.066966 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" event={"ID":"5b17231c-1edb-4639-8333-678e3434f4dd","Type":"ContainerDied","Data":"60a1142fe263841ea85847e103545520b46aee8733cb982cb5e8bd7b8fc70379"} Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.069391 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-77v4n" event={"ID":"aad96861-15f8-4715-8b7c-927ae48fe59c","Type":"ContainerStarted","Data":"2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a"} Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.080708 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.161430 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.360229 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gs786-config-j9dtl"] Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.361810 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.365308 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.375146 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gs786-config-j9dtl"] Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.382722 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.383008 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.383104 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.383296 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.383429 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4slr\" (UniqueName: \"kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.383561 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.485535 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.485883 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.485907 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486072 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486162 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486363 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486444 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4slr\" (UniqueName: \"kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486448 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.486401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.487120 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.488168 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.504051 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4slr\" (UniqueName: \"kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr\") pod \"ovn-controller-gs786-config-j9dtl\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.678074 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:05 crc kubenswrapper[5018]: I1014 07:08:05.970601 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gs786-config-j9dtl"] Oct 14 07:08:05 crc kubenswrapper[5018]: W1014 07:08:05.981458 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27385486_c96d_4cd7_8d69_e9cc283535ff.slice/crio-04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92 WatchSource:0}: Error finding container 04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92: Status 404 returned error can't find the container with id 04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92 Oct 14 07:08:06 crc kubenswrapper[5018]: I1014 07:08:06.095513 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786-config-j9dtl" event={"ID":"27385486-c96d-4cd7-8d69-e9cc283535ff","Type":"ContainerStarted","Data":"04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92"} Oct 14 07:08:06 crc kubenswrapper[5018]: I1014 07:08:06.100640 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" event={"ID":"5b17231c-1edb-4639-8333-678e3434f4dd","Type":"ContainerStarted","Data":"f4e9706118e43b2c9cdcc673466db21989bfa65b18943fa1cbdd21b5134b0bc7"} Oct 14 07:08:06 crc kubenswrapper[5018]: I1014 07:08:06.100760 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:06 crc kubenswrapper[5018]: I1014 07:08:06.120667 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" podStartSLOduration=3.120643056 podStartE2EDuration="3.120643056s" podCreationTimestamp="2025-10-14 07:08:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:06.118432105 +0000 UTC m=+1102.702478722" watchObservedRunningTime="2025-10-14 07:08:06.120643056 +0000 UTC m=+1102.704689673" Oct 14 07:08:07 crc kubenswrapper[5018]: I1014 07:08:07.108765 5018 generic.go:334] "Generic (PLEG): container finished" podID="27385486-c96d-4cd7-8d69-e9cc283535ff" containerID="1be68f6f0cba15c3bc66c45d56c706fc1f7a4a7231a6283e04c05aaac90e6232" exitCode=0 Oct 14 07:08:07 crc kubenswrapper[5018]: I1014 07:08:07.108851 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786-config-j9dtl" event={"ID":"27385486-c96d-4cd7-8d69-e9cc283535ff","Type":"ContainerDied","Data":"1be68f6f0cba15c3bc66c45d56c706fc1f7a4a7231a6283e04c05aaac90e6232"} Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.423812 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532077 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4slr\" (UniqueName: \"kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532125 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532259 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532360 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532504 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts\") pod \"27385486-c96d-4cd7-8d69-e9cc283535ff\" (UID: \"27385486-c96d-4cd7-8d69-e9cc283535ff\") " Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532770 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532818 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.532846 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run" (OuterVolumeSpecName: "var-run") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.533136 5018 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.533155 5018 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.533164 5018 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/27385486-c96d-4cd7-8d69-e9cc283535ff-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.533520 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.533779 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts" (OuterVolumeSpecName: "scripts") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.539004 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr" (OuterVolumeSpecName: "kube-api-access-b4slr") pod "27385486-c96d-4cd7-8d69-e9cc283535ff" (UID: "27385486-c96d-4cd7-8d69-e9cc283535ff"). InnerVolumeSpecName "kube-api-access-b4slr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.634368 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.634396 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4slr\" (UniqueName: \"kubernetes.io/projected/27385486-c96d-4cd7-8d69-e9cc283535ff-kube-api-access-b4slr\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:08 crc kubenswrapper[5018]: I1014 07:08:08.634413 5018 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/27385486-c96d-4cd7-8d69-e9cc283535ff-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:09 crc kubenswrapper[5018]: I1014 07:08:09.133076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786-config-j9dtl" event={"ID":"27385486-c96d-4cd7-8d69-e9cc283535ff","Type":"ContainerDied","Data":"04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92"} Oct 14 07:08:09 crc kubenswrapper[5018]: I1014 07:08:09.133118 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04700de931281d85a4956b70b09ea3051b910090744eb68e944474b50cbfaf92" Oct 14 07:08:09 crc kubenswrapper[5018]: I1014 07:08:09.133169 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786-config-j9dtl" Oct 14 07:08:09 crc kubenswrapper[5018]: I1014 07:08:09.538660 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gs786-config-j9dtl"] Oct 14 07:08:09 crc kubenswrapper[5018]: I1014 07:08:09.545642 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gs786-config-j9dtl"] Oct 14 07:08:10 crc kubenswrapper[5018]: I1014 07:08:10.017889 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gs786" Oct 14 07:08:10 crc kubenswrapper[5018]: I1014 07:08:10.643802 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27385486-c96d-4cd7-8d69-e9cc283535ff" path="/var/lib/kubelet/pods/27385486-c96d-4cd7-8d69-e9cc283535ff/volumes" Oct 14 07:08:13 crc kubenswrapper[5018]: I1014 07:08:13.682101 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:13 crc kubenswrapper[5018]: I1014 07:08:13.786865 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:08:13 crc kubenswrapper[5018]: I1014 07:08:13.787128 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" containerID="cri-o://f94a6340ff419e5609d0db3cdb8fcc05fdb13312af60401fe1ada92fd6712189" gracePeriod=10 Oct 14 07:08:14 crc kubenswrapper[5018]: I1014 07:08:14.184228 5018 generic.go:334] "Generic (PLEG): container finished" podID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerID="f94a6340ff419e5609d0db3cdb8fcc05fdb13312af60401fe1ada92fd6712189" exitCode=0 Oct 14 07:08:14 crc kubenswrapper[5018]: I1014 07:08:14.184272 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" event={"ID":"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad","Type":"ContainerDied","Data":"f94a6340ff419e5609d0db3cdb8fcc05fdb13312af60401fe1ada92fd6712189"} Oct 14 07:08:15 crc kubenswrapper[5018]: I1014 07:08:15.618508 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 14 07:08:20 crc kubenswrapper[5018]: I1014 07:08:20.313913 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:08:20 crc kubenswrapper[5018]: I1014 07:08:20.354966 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 07:08:20 crc kubenswrapper[5018]: I1014 07:08:20.618993 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 14 07:08:21 crc kubenswrapper[5018]: E1014 07:08:21.281805 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:839f0e593dd6b59e385ec9471f4eeaa34f1c539268588114cbc34cc9a6117835" Oct 14 07:08:21 crc kubenswrapper[5018]: E1014 07:08:21.282049 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:839f0e593dd6b59e385ec9471f4eeaa34f1c539268588114cbc34cc9a6117835,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxkxd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-77v4n_openstack(aad96861-15f8-4715-8b7c-927ae48fe59c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:08:21 crc kubenswrapper[5018]: E1014 07:08:21.283383 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-77v4n" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.729092 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.817207 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config\") pod \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.817286 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb\") pod \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.817342 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc\") pod \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.817486 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnxr4\" (UniqueName: \"kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4\") pod \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.817524 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb\") pod \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\" (UID: \"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad\") " Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.884120 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4" (OuterVolumeSpecName: "kube-api-access-tnxr4") pod "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" (UID: "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad"). InnerVolumeSpecName "kube-api-access-tnxr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.892474 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config" (OuterVolumeSpecName: "config") pod "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" (UID: "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.916215 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" (UID: "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.917117 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" (UID: "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.919254 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.919348 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.919362 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.919375 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnxr4\" (UniqueName: \"kubernetes.io/projected/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-kube-api-access-tnxr4\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:21 crc kubenswrapper[5018]: I1014 07:08:21.925639 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" (UID: "57f1b1b7-2e2d-49a9-822f-8ffbd55251ad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.020337 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.210873 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6mb75"] Oct 14 07:08:22 crc kubenswrapper[5018]: E1014 07:08:22.211449 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.211463 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" Oct 14 07:08:22 crc kubenswrapper[5018]: E1014 07:08:22.211479 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27385486-c96d-4cd7-8d69-e9cc283535ff" containerName="ovn-config" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.211485 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="27385486-c96d-4cd7-8d69-e9cc283535ff" containerName="ovn-config" Oct 14 07:08:22 crc kubenswrapper[5018]: E1014 07:08:22.211504 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="init" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.211510 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="init" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.211678 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="27385486-c96d-4cd7-8d69-e9cc283535ff" containerName="ovn-config" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.211694 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" containerName="dnsmasq-dns" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.212199 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.272142 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" event={"ID":"57f1b1b7-2e2d-49a9-822f-8ffbd55251ad","Type":"ContainerDied","Data":"2902f42e015becb2b9213efcf9bcf9bff6186fcadda218b81eba4aa14f10a327"} Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.272418 5018 scope.go:117] "RemoveContainer" containerID="f94a6340ff419e5609d0db3cdb8fcc05fdb13312af60401fe1ada92fd6712189" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.272205 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c757dd68f-whbm8" Oct 14 07:08:22 crc kubenswrapper[5018]: E1014 07:08:22.274263 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:839f0e593dd6b59e385ec9471f4eeaa34f1c539268588114cbc34cc9a6117835\\\"\"" pod="openstack/glance-db-sync-77v4n" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.276128 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6mb75"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.293979 5018 scope.go:117] "RemoveContainer" containerID="cd4b9f95ca796b2e388b05c10f01ec12a79b441d49ce881004a2d675acfb271d" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.324586 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t976j\" (UniqueName: \"kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j\") pod \"cinder-db-create-6mb75\" (UID: \"652c0172-b19f-4bfa-9f82-3ad458cbc8cd\") " pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.335564 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lrzn8"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.336687 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.347334 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.351827 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c757dd68f-whbm8"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.356692 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lrzn8"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.426169 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p7tt\" (UniqueName: \"kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt\") pod \"barbican-db-create-lrzn8\" (UID: \"ba453765-5fa8-4930-b9af-155c98c87d01\") " pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.426570 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t976j\" (UniqueName: \"kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j\") pod \"cinder-db-create-6mb75\" (UID: \"652c0172-b19f-4bfa-9f82-3ad458cbc8cd\") " pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.444086 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t976j\" (UniqueName: \"kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j\") pod \"cinder-db-create-6mb75\" (UID: \"652c0172-b19f-4bfa-9f82-3ad458cbc8cd\") " pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.512118 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fzlkk"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.513373 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.520525 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fzlkk"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.546326 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.546824 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p7tt\" (UniqueName: \"kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt\") pod \"barbican-db-create-lrzn8\" (UID: \"ba453765-5fa8-4930-b9af-155c98c87d01\") " pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.597256 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-hhwdt"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.598288 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p7tt\" (UniqueName: \"kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt\") pod \"barbican-db-create-lrzn8\" (UID: \"ba453765-5fa8-4930-b9af-155c98c87d01\") " pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.598729 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.603152 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.603206 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.603163 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.603550 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ptvv" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.620390 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f1b1b7-2e2d-49a9-822f-8ffbd55251ad" path="/var/lib/kubelet/pods/57f1b1b7-2e2d-49a9-822f-8ffbd55251ad/volumes" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.625089 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hhwdt"] Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.648371 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27fx9\" (UniqueName: \"kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9\") pod \"neutron-db-create-fzlkk\" (UID: \"33a935d4-85ef-44c2-967b-6dbd9b7c95d8\") " pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.666820 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.749718 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27fx9\" (UniqueName: \"kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9\") pod \"neutron-db-create-fzlkk\" (UID: \"33a935d4-85ef-44c2-967b-6dbd9b7c95d8\") " pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.749787 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.749810 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vmqz\" (UniqueName: \"kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.749853 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.766282 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27fx9\" (UniqueName: \"kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9\") pod \"neutron-db-create-fzlkk\" (UID: \"33a935d4-85ef-44c2-967b-6dbd9b7c95d8\") " pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.851751 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.851790 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vmqz\" (UniqueName: \"kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.851835 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.856530 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.857343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.868629 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.874278 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vmqz\" (UniqueName: \"kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz\") pod \"keystone-db-sync-hhwdt\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.921302 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lrzn8"] Oct 14 07:08:22 crc kubenswrapper[5018]: W1014 07:08:22.923069 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba453765_5fa8_4930_b9af_155c98c87d01.slice/crio-00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c WatchSource:0}: Error finding container 00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c: Status 404 returned error can't find the container with id 00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c Oct 14 07:08:22 crc kubenswrapper[5018]: I1014 07:08:22.956522 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.022133 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6mb75"] Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.269408 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hhwdt"] Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.283991 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6mb75" event={"ID":"652c0172-b19f-4bfa-9f82-3ad458cbc8cd","Type":"ContainerStarted","Data":"fe9fc3a7516d66299c812a398eed4f6171bc4742bb3abe2c90dde2b7dc5c537a"} Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.287709 5018 generic.go:334] "Generic (PLEG): container finished" podID="ba453765-5fa8-4930-b9af-155c98c87d01" containerID="9e8038bd5aa84b2d30b0819224bdfdd25f526085072462f8308583a788feaba3" exitCode=0 Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.287733 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lrzn8" event={"ID":"ba453765-5fa8-4930-b9af-155c98c87d01","Type":"ContainerDied","Data":"9e8038bd5aa84b2d30b0819224bdfdd25f526085072462f8308583a788feaba3"} Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.287750 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lrzn8" event={"ID":"ba453765-5fa8-4930-b9af-155c98c87d01","Type":"ContainerStarted","Data":"00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c"} Oct 14 07:08:23 crc kubenswrapper[5018]: I1014 07:08:23.380655 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fzlkk"] Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.305597 5018 generic.go:334] "Generic (PLEG): container finished" podID="33a935d4-85ef-44c2-967b-6dbd9b7c95d8" containerID="b310644847b9e57cc0725838b5a8468110e9a5e7031ec226076e90392203de20" exitCode=0 Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.305794 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzlkk" event={"ID":"33a935d4-85ef-44c2-967b-6dbd9b7c95d8","Type":"ContainerDied","Data":"b310644847b9e57cc0725838b5a8468110e9a5e7031ec226076e90392203de20"} Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.306132 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzlkk" event={"ID":"33a935d4-85ef-44c2-967b-6dbd9b7c95d8","Type":"ContainerStarted","Data":"eff0c11257aba3e241ce39991e62894b375b8c7234ad97c5fbdc07d59d5869b9"} Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.309317 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hhwdt" event={"ID":"4c45b3a9-3966-4584-ac73-138eed247f9d","Type":"ContainerStarted","Data":"c9fb04f0876db10cd902074603671c49287eae3df7487e6dd23f567fcdc3624a"} Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.312803 5018 generic.go:334] "Generic (PLEG): container finished" podID="652c0172-b19f-4bfa-9f82-3ad458cbc8cd" containerID="9d51216aaee8d1ff3d268268f6c176157e2126998ecf32f4f13de3c851aac61f" exitCode=0 Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.313083 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6mb75" event={"ID":"652c0172-b19f-4bfa-9f82-3ad458cbc8cd","Type":"ContainerDied","Data":"9d51216aaee8d1ff3d268268f6c176157e2126998ecf32f4f13de3c851aac61f"} Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.667784 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.786862 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p7tt\" (UniqueName: \"kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt\") pod \"ba453765-5fa8-4930-b9af-155c98c87d01\" (UID: \"ba453765-5fa8-4930-b9af-155c98c87d01\") " Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.794336 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt" (OuterVolumeSpecName: "kube-api-access-2p7tt") pod "ba453765-5fa8-4930-b9af-155c98c87d01" (UID: "ba453765-5fa8-4930-b9af-155c98c87d01"). InnerVolumeSpecName "kube-api-access-2p7tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:24 crc kubenswrapper[5018]: I1014 07:08:24.889126 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p7tt\" (UniqueName: \"kubernetes.io/projected/ba453765-5fa8-4930-b9af-155c98c87d01-kube-api-access-2p7tt\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:25 crc kubenswrapper[5018]: I1014 07:08:25.322978 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lrzn8" Oct 14 07:08:25 crc kubenswrapper[5018]: I1014 07:08:25.323039 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lrzn8" event={"ID":"ba453765-5fa8-4930-b9af-155c98c87d01","Type":"ContainerDied","Data":"00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c"} Oct 14 07:08:25 crc kubenswrapper[5018]: I1014 07:08:25.323084 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00f93787c3cdbfc8471ca7fc7007f115cd6cd3a91afec1572893ca7b44be847c" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.367832 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.375405 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.385192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fzlkk" event={"ID":"33a935d4-85ef-44c2-967b-6dbd9b7c95d8","Type":"ContainerDied","Data":"eff0c11257aba3e241ce39991e62894b375b8c7234ad97c5fbdc07d59d5869b9"} Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.385237 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eff0c11257aba3e241ce39991e62894b375b8c7234ad97c5fbdc07d59d5869b9" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.385284 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fzlkk" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.390588 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6mb75" event={"ID":"652c0172-b19f-4bfa-9f82-3ad458cbc8cd","Type":"ContainerDied","Data":"fe9fc3a7516d66299c812a398eed4f6171bc4742bb3abe2c90dde2b7dc5c537a"} Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.390646 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6mb75" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.390656 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe9fc3a7516d66299c812a398eed4f6171bc4742bb3abe2c90dde2b7dc5c537a" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.554873 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t976j\" (UniqueName: \"kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j\") pod \"652c0172-b19f-4bfa-9f82-3ad458cbc8cd\" (UID: \"652c0172-b19f-4bfa-9f82-3ad458cbc8cd\") " Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.555181 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27fx9\" (UniqueName: \"kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9\") pod \"33a935d4-85ef-44c2-967b-6dbd9b7c95d8\" (UID: \"33a935d4-85ef-44c2-967b-6dbd9b7c95d8\") " Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.562370 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j" (OuterVolumeSpecName: "kube-api-access-t976j") pod "652c0172-b19f-4bfa-9f82-3ad458cbc8cd" (UID: "652c0172-b19f-4bfa-9f82-3ad458cbc8cd"). InnerVolumeSpecName "kube-api-access-t976j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.564204 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9" (OuterVolumeSpecName: "kube-api-access-27fx9") pod "33a935d4-85ef-44c2-967b-6dbd9b7c95d8" (UID: "33a935d4-85ef-44c2-967b-6dbd9b7c95d8"). InnerVolumeSpecName "kube-api-access-27fx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.657490 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t976j\" (UniqueName: \"kubernetes.io/projected/652c0172-b19f-4bfa-9f82-3ad458cbc8cd-kube-api-access-t976j\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:28 crc kubenswrapper[5018]: I1014 07:08:28.657532 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27fx9\" (UniqueName: \"kubernetes.io/projected/33a935d4-85ef-44c2-967b-6dbd9b7c95d8-kube-api-access-27fx9\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:29 crc kubenswrapper[5018]: I1014 07:08:29.400758 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hhwdt" event={"ID":"4c45b3a9-3966-4584-ac73-138eed247f9d","Type":"ContainerStarted","Data":"1a0f3efdf0f6304ed75d79e70ae3f3de5dbdbc727d01370fd2c7e00dba266640"} Oct 14 07:08:29 crc kubenswrapper[5018]: I1014 07:08:29.424988 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-hhwdt" podStartSLOduration=2.55376162 podStartE2EDuration="7.424968329s" podCreationTimestamp="2025-10-14 07:08:22 +0000 UTC" firstStartedPulling="2025-10-14 07:08:23.360582132 +0000 UTC m=+1119.944628769" lastFinishedPulling="2025-10-14 07:08:28.231788821 +0000 UTC m=+1124.815835478" observedRunningTime="2025-10-14 07:08:29.423405176 +0000 UTC m=+1126.007451803" watchObservedRunningTime="2025-10-14 07:08:29.424968329 +0000 UTC m=+1126.009014956" Oct 14 07:08:31 crc kubenswrapper[5018]: I1014 07:08:31.423395 5018 generic.go:334] "Generic (PLEG): container finished" podID="4c45b3a9-3966-4584-ac73-138eed247f9d" containerID="1a0f3efdf0f6304ed75d79e70ae3f3de5dbdbc727d01370fd2c7e00dba266640" exitCode=0 Oct 14 07:08:31 crc kubenswrapper[5018]: I1014 07:08:31.423578 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hhwdt" event={"ID":"4c45b3a9-3966-4584-ac73-138eed247f9d","Type":"ContainerDied","Data":"1a0f3efdf0f6304ed75d79e70ae3f3de5dbdbc727d01370fd2c7e00dba266640"} Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.377437 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0076-account-create-pdx2z"] Oct 14 07:08:32 crc kubenswrapper[5018]: E1014 07:08:32.378040 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652c0172-b19f-4bfa-9f82-3ad458cbc8cd" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378059 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="652c0172-b19f-4bfa-9f82-3ad458cbc8cd" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: E1014 07:08:32.378111 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba453765-5fa8-4930-b9af-155c98c87d01" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378122 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba453765-5fa8-4930-b9af-155c98c87d01" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: E1014 07:08:32.378136 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a935d4-85ef-44c2-967b-6dbd9b7c95d8" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378157 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a935d4-85ef-44c2-967b-6dbd9b7c95d8" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378433 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="652c0172-b19f-4bfa-9f82-3ad458cbc8cd" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378459 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a935d4-85ef-44c2-967b-6dbd9b7c95d8" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.378479 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba453765-5fa8-4930-b9af-155c98c87d01" containerName="mariadb-database-create" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.379317 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.382478 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.384171 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0076-account-create-pdx2z"] Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.463822 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.463875 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.550500 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5769-account-create-487kz"] Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.551998 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.554145 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.560125 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5769-account-create-487kz"] Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.566061 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5gz\" (UniqueName: \"kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz\") pod \"barbican-0076-account-create-pdx2z\" (UID: \"f76f5af3-9474-4186-9da0-d5ea4fe14f09\") " pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.670581 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5gz\" (UniqueName: \"kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz\") pod \"barbican-0076-account-create-pdx2z\" (UID: \"f76f5af3-9474-4186-9da0-d5ea4fe14f09\") " pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.670732 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtt56\" (UniqueName: \"kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56\") pod \"cinder-5769-account-create-487kz\" (UID: \"e474ca2f-3868-45b7-9251-4e254c8c14dd\") " pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.688361 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5gz\" (UniqueName: \"kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz\") pod \"barbican-0076-account-create-pdx2z\" (UID: \"f76f5af3-9474-4186-9da0-d5ea4fe14f09\") " pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.715083 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.772530 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtt56\" (UniqueName: \"kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56\") pod \"cinder-5769-account-create-487kz\" (UID: \"e474ca2f-3868-45b7-9251-4e254c8c14dd\") " pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.781306 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.803996 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtt56\" (UniqueName: \"kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56\") pod \"cinder-5769-account-create-487kz\" (UID: \"e474ca2f-3868-45b7-9251-4e254c8c14dd\") " pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.871253 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.975444 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle\") pod \"4c45b3a9-3966-4584-ac73-138eed247f9d\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.975507 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vmqz\" (UniqueName: \"kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz\") pod \"4c45b3a9-3966-4584-ac73-138eed247f9d\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.975595 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data\") pod \"4c45b3a9-3966-4584-ac73-138eed247f9d\" (UID: \"4c45b3a9-3966-4584-ac73-138eed247f9d\") " Oct 14 07:08:32 crc kubenswrapper[5018]: I1014 07:08:32.979761 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz" (OuterVolumeSpecName: "kube-api-access-5vmqz") pod "4c45b3a9-3966-4584-ac73-138eed247f9d" (UID: "4c45b3a9-3966-4584-ac73-138eed247f9d"). InnerVolumeSpecName "kube-api-access-5vmqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.010153 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c45b3a9-3966-4584-ac73-138eed247f9d" (UID: "4c45b3a9-3966-4584-ac73-138eed247f9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.025199 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data" (OuterVolumeSpecName: "config-data") pod "4c45b3a9-3966-4584-ac73-138eed247f9d" (UID: "4c45b3a9-3966-4584-ac73-138eed247f9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.079413 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.079465 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vmqz\" (UniqueName: \"kubernetes.io/projected/4c45b3a9-3966-4584-ac73-138eed247f9d-kube-api-access-5vmqz\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.079480 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c45b3a9-3966-4584-ac73-138eed247f9d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.086446 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5769-account-create-487kz"] Oct 14 07:08:33 crc kubenswrapper[5018]: W1014 07:08:33.092838 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode474ca2f_3868_45b7_9251_4e254c8c14dd.slice/crio-7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85 WatchSource:0}: Error finding container 7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85: Status 404 returned error can't find the container with id 7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85 Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.148364 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0076-account-create-pdx2z"] Oct 14 07:08:33 crc kubenswrapper[5018]: W1014 07:08:33.160817 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf76f5af3_9474_4186_9da0_d5ea4fe14f09.slice/crio-23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694 WatchSource:0}: Error finding container 23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694: Status 404 returned error can't find the container with id 23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694 Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.443036 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hhwdt" event={"ID":"4c45b3a9-3966-4584-ac73-138eed247f9d","Type":"ContainerDied","Data":"c9fb04f0876db10cd902074603671c49287eae3df7487e6dd23f567fcdc3624a"} Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.443133 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9fb04f0876db10cd902074603671c49287eae3df7487e6dd23f567fcdc3624a" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.443077 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hhwdt" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.444917 5018 generic.go:334] "Generic (PLEG): container finished" podID="f76f5af3-9474-4186-9da0-d5ea4fe14f09" containerID="89ca651abd56061b3d4f4730fa92495f37b9b76c796d8b6fdf496276d6c696c2" exitCode=0 Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.444993 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0076-account-create-pdx2z" event={"ID":"f76f5af3-9474-4186-9da0-d5ea4fe14f09","Type":"ContainerDied","Data":"89ca651abd56061b3d4f4730fa92495f37b9b76c796d8b6fdf496276d6c696c2"} Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.445036 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0076-account-create-pdx2z" event={"ID":"f76f5af3-9474-4186-9da0-d5ea4fe14f09","Type":"ContainerStarted","Data":"23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694"} Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.446658 5018 generic.go:334] "Generic (PLEG): container finished" podID="e474ca2f-3868-45b7-9251-4e254c8c14dd" containerID="855cef651e93fc582bf4dfce02cadbb1a3d709865a3fde078f280d5e771dc195" exitCode=0 Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.446721 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5769-account-create-487kz" event={"ID":"e474ca2f-3868-45b7-9251-4e254c8c14dd","Type":"ContainerDied","Data":"855cef651e93fc582bf4dfce02cadbb1a3d709865a3fde078f280d5e771dc195"} Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.446775 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5769-account-create-487kz" event={"ID":"e474ca2f-3868-45b7-9251-4e254c8c14dd","Type":"ContainerStarted","Data":"7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85"} Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.733050 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:33 crc kubenswrapper[5018]: E1014 07:08:33.733675 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c45b3a9-3966-4584-ac73-138eed247f9d" containerName="keystone-db-sync" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.733688 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c45b3a9-3966-4584-ac73-138eed247f9d" containerName="keystone-db-sync" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.733876 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c45b3a9-3966-4584-ac73-138eed247f9d" containerName="keystone-db-sync" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.736890 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.757417 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.763838 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-r8t2m"] Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.764813 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.766148 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.766200 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.766687 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.767878 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ptvv" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.778877 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r8t2m"] Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896369 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896428 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896458 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqzs5\" (UniqueName: \"kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896476 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896544 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896567 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896607 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hk9\" (UniqueName: \"kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896644 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896664 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896691 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896716 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.896738 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.936876 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.964833 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.970055 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.970363 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999353 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hk9\" (UniqueName: \"kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999394 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8rsx\" (UniqueName: \"kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999415 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999437 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999454 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999472 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999490 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999505 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999519 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999540 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999560 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999582 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999605 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999629 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999661 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqzs5\" (UniqueName: \"kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999678 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999710 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:33 crc kubenswrapper[5018]: I1014 07:08:33.999728 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:33.999748 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.000530 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.002443 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.003141 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.004281 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.004407 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.005060 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.015608 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.015755 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.016250 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.016699 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.019217 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqzs5\" (UniqueName: \"kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.022077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hk9\" (UniqueName: \"kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9\") pod \"dnsmasq-dns-b6dcd555f-v4dvl\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.022367 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data\") pod \"keystone-bootstrap-r8t2m\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.040029 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.040572 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.069778 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4nzkd"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.070898 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.077733 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vpkpb" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.077926 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.078430 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.081897 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.082323 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.085882 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.100438 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4nzkd"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.110247 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116398 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116530 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8rsx\" (UniqueName: \"kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116584 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116622 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116666 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116701 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116725 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.116812 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.117244 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.117282 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxc99\" (UniqueName: \"kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.117325 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.117418 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.119956 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.125878 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.126363 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.130988 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.131470 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.134487 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.152385 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8rsx\" (UniqueName: \"kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx\") pod \"ceilometer-0\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.222560 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.222904 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58g7k\" (UniqueName: \"kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.222948 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.222967 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.222986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxc99\" (UniqueName: \"kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223008 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223024 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223049 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223076 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223096 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.223140 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.232082 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.232173 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.232347 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.232747 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.244217 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxc99\" (UniqueName: \"kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99\") pod \"placement-db-sync-4nzkd\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.300068 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.328073 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.328990 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.329171 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332010 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332182 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332345 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58g7k\" (UniqueName: \"kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332466 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332660 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.332824 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.333552 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.336542 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.351085 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58g7k\" (UniqueName: \"kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k\") pod \"dnsmasq-dns-5c75bdf7bf-x9gtp\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.479076 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.485818 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.525135 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.568726 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:08:34 crc kubenswrapper[5018]: W1014 07:08:34.575578 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88ef1413_475d_4f9b_ac37_24b52ed55ec8.slice/crio-45345be79451f5f7914a23a88d01e66b4f0c4f096c22c6dd1c9a33055746c20e WatchSource:0}: Error finding container 45345be79451f5f7914a23a88d01e66b4f0c4f096c22c6dd1c9a33055746c20e: Status 404 returned error can't find the container with id 45345be79451f5f7914a23a88d01e66b4f0c4f096c22c6dd1c9a33055746c20e Oct 14 07:08:34 crc kubenswrapper[5018]: W1014 07:08:34.578182 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2af870e0_db4f_49ba_afeb_238282c56c61.slice/crio-80121411d168fc7e2f3e52c9ebe3811af90e9eb7ba2616ad2409acf0eb39acc7 WatchSource:0}: Error finding container 80121411d168fc7e2f3e52c9ebe3811af90e9eb7ba2616ad2409acf0eb39acc7: Status 404 returned error can't find the container with id 80121411d168fc7e2f3e52c9ebe3811af90e9eb7ba2616ad2409acf0eb39acc7 Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.628675 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r8t2m"] Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.906486 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.911516 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.949573 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b5gz\" (UniqueName: \"kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz\") pod \"f76f5af3-9474-4186-9da0-d5ea4fe14f09\" (UID: \"f76f5af3-9474-4186-9da0-d5ea4fe14f09\") " Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.949783 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtt56\" (UniqueName: \"kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56\") pod \"e474ca2f-3868-45b7-9251-4e254c8c14dd\" (UID: \"e474ca2f-3868-45b7-9251-4e254c8c14dd\") " Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.959132 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56" (OuterVolumeSpecName: "kube-api-access-gtt56") pod "e474ca2f-3868-45b7-9251-4e254c8c14dd" (UID: "e474ca2f-3868-45b7-9251-4e254c8c14dd"). InnerVolumeSpecName "kube-api-access-gtt56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:34 crc kubenswrapper[5018]: I1014 07:08:34.959220 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz" (OuterVolumeSpecName: "kube-api-access-7b5gz") pod "f76f5af3-9474-4186-9da0-d5ea4fe14f09" (UID: "f76f5af3-9474-4186-9da0-d5ea4fe14f09"). InnerVolumeSpecName "kube-api-access-7b5gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.024302 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4nzkd"] Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.051826 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtt56\" (UniqueName: \"kubernetes.io/projected/e474ca2f-3868-45b7-9251-4e254c8c14dd-kube-api-access-gtt56\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.051868 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b5gz\" (UniqueName: \"kubernetes.io/projected/f76f5af3-9474-4186-9da0-d5ea4fe14f09-kube-api-access-7b5gz\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:35 crc kubenswrapper[5018]: W1014 07:08:35.084820 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6920d271_45c6_4d58_af8a_6707e5fa80e9.slice/crio-ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332 WatchSource:0}: Error finding container ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332: Status 404 returned error can't find the container with id ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332 Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.188930 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:08:35 crc kubenswrapper[5018]: W1014 07:08:35.198174 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8786087b_cf37_429a_9c1b_bd057f422c3d.slice/crio-d68f2454841be95f2ee71cceb8da693e91928c0be731adcd3ffe6f6b6779a6cd WatchSource:0}: Error finding container d68f2454841be95f2ee71cceb8da693e91928c0be731adcd3ffe6f6b6779a6cd: Status 404 returned error can't find the container with id d68f2454841be95f2ee71cceb8da693e91928c0be731adcd3ffe6f6b6779a6cd Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.467337 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5769-account-create-487kz" event={"ID":"e474ca2f-3868-45b7-9251-4e254c8c14dd","Type":"ContainerDied","Data":"7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.467399 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c56eb67a78a4f4eed0121c13d447f5cb156297848be21a1ed0f9da9d2335d85" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.467488 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5769-account-create-487kz" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.470733 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4nzkd" event={"ID":"6920d271-45c6-4d58-af8a-6707e5fa80e9","Type":"ContainerStarted","Data":"ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.472821 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r8t2m" event={"ID":"456702b4-2d47-4803-b768-ebc9dcc679eb","Type":"ContainerStarted","Data":"eb63bc0308d56dec44abe934bec14d642df71515d6ac71e9bf8281b994bb49a4"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.472866 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r8t2m" event={"ID":"456702b4-2d47-4803-b768-ebc9dcc679eb","Type":"ContainerStarted","Data":"086044733bdef854a2cd11ab51ecbae5380d66c69118eb362c3e83c01e15eb6a"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.475424 5018 generic.go:334] "Generic (PLEG): container finished" podID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerID="3a3ea19ac068e25ad6df8b649097f8de9ef2b573ca06c17189eba6c123e13c2b" exitCode=0 Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.475500 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" event={"ID":"8786087b-cf37-429a-9c1b-bd057f422c3d","Type":"ContainerDied","Data":"3a3ea19ac068e25ad6df8b649097f8de9ef2b573ca06c17189eba6c123e13c2b"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.475528 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" event={"ID":"8786087b-cf37-429a-9c1b-bd057f422c3d","Type":"ContainerStarted","Data":"d68f2454841be95f2ee71cceb8da693e91928c0be731adcd3ffe6f6b6779a6cd"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.487286 5018 generic.go:334] "Generic (PLEG): container finished" podID="88ef1413-475d-4f9b-ac37-24b52ed55ec8" containerID="f74332e94e043f84d8d522667ea5a345f1b92dc94fd21aded986f06e4284a0a0" exitCode=0 Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.487334 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" event={"ID":"88ef1413-475d-4f9b-ac37-24b52ed55ec8","Type":"ContainerDied","Data":"f74332e94e043f84d8d522667ea5a345f1b92dc94fd21aded986f06e4284a0a0"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.487387 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" event={"ID":"88ef1413-475d-4f9b-ac37-24b52ed55ec8","Type":"ContainerStarted","Data":"45345be79451f5f7914a23a88d01e66b4f0c4f096c22c6dd1c9a33055746c20e"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.499399 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0076-account-create-pdx2z" event={"ID":"f76f5af3-9474-4186-9da0-d5ea4fe14f09","Type":"ContainerDied","Data":"23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.499434 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23f63ccbf23ae480d10dc6ec9399d7920cc0fa45383e86babd41ab56e0490694" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.499478 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0076-account-create-pdx2z" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.501592 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerStarted","Data":"80121411d168fc7e2f3e52c9ebe3811af90e9eb7ba2616ad2409acf0eb39acc7"} Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.502810 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-r8t2m" podStartSLOduration=2.502797405 podStartE2EDuration="2.502797405s" podCreationTimestamp="2025-10-14 07:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:35.4927983 +0000 UTC m=+1132.076844927" watchObservedRunningTime="2025-10-14 07:08:35.502797405 +0000 UTC m=+1132.086844032" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.842396 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.865702 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969303 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969357 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969394 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969440 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hk9\" (UniqueName: \"kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.969468 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb\") pod \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\" (UID: \"88ef1413-475d-4f9b-ac37-24b52ed55ec8\") " Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.988101 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9" (OuterVolumeSpecName: "kube-api-access-h7hk9") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "kube-api-access-h7hk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.991696 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.999144 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:35 crc kubenswrapper[5018]: I1014 07:08:35.999641 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.006468 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config" (OuterVolumeSpecName: "config") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.011366 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "88ef1413-475d-4f9b-ac37-24b52ed55ec8" (UID: "88ef1413-475d-4f9b-ac37-24b52ed55ec8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.070899 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.070953 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.070966 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.070979 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.070991 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hk9\" (UniqueName: \"kubernetes.io/projected/88ef1413-475d-4f9b-ac37-24b52ed55ec8-kube-api-access-h7hk9\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.071000 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88ef1413-475d-4f9b-ac37-24b52ed55ec8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.514522 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" event={"ID":"8786087b-cf37-429a-9c1b-bd057f422c3d","Type":"ContainerStarted","Data":"9df84315e41ea8b5a5e789dd8ace05732ab56f75aebfeec16c62b24045e65329"} Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.514669 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.518795 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" event={"ID":"88ef1413-475d-4f9b-ac37-24b52ed55ec8","Type":"ContainerDied","Data":"45345be79451f5f7914a23a88d01e66b4f0c4f096c22c6dd1c9a33055746c20e"} Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.518845 5018 scope.go:117] "RemoveContainer" containerID="f74332e94e043f84d8d522667ea5a345f1b92dc94fd21aded986f06e4284a0a0" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.518947 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6dcd555f-v4dvl" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.524712 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-77v4n" event={"ID":"aad96861-15f8-4715-8b7c-927ae48fe59c","Type":"ContainerStarted","Data":"a76b17433cbbe67cc02ee352389554f84d8ac0be588c5e00e58f7115928d5574"} Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.543895 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" podStartSLOduration=2.5438773059999997 podStartE2EDuration="2.543877306s" podCreationTimestamp="2025-10-14 07:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:36.542120428 +0000 UTC m=+1133.126167055" watchObservedRunningTime="2025-10-14 07:08:36.543877306 +0000 UTC m=+1133.127923933" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.563773 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-77v4n" podStartSLOduration=3.206039417 podStartE2EDuration="33.563750933s" podCreationTimestamp="2025-10-14 07:08:03 +0000 UTC" firstStartedPulling="2025-10-14 07:08:04.779426685 +0000 UTC m=+1101.363473322" lastFinishedPulling="2025-10-14 07:08:35.137138211 +0000 UTC m=+1131.721184838" observedRunningTime="2025-10-14 07:08:36.561113531 +0000 UTC m=+1133.145160158" watchObservedRunningTime="2025-10-14 07:08:36.563750933 +0000 UTC m=+1133.147797570" Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.616261 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:36 crc kubenswrapper[5018]: I1014 07:08:36.621158 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6dcd555f-v4dvl"] Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667048 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pnqtc"] Oct 14 07:08:37 crc kubenswrapper[5018]: E1014 07:08:37.667338 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ef1413-475d-4f9b-ac37-24b52ed55ec8" containerName="init" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667351 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ef1413-475d-4f9b-ac37-24b52ed55ec8" containerName="init" Oct 14 07:08:37 crc kubenswrapper[5018]: E1014 07:08:37.667362 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76f5af3-9474-4186-9da0-d5ea4fe14f09" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667368 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76f5af3-9474-4186-9da0-d5ea4fe14f09" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: E1014 07:08:37.667395 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e474ca2f-3868-45b7-9251-4e254c8c14dd" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667401 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e474ca2f-3868-45b7-9251-4e254c8c14dd" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667560 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e474ca2f-3868-45b7-9251-4e254c8c14dd" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667571 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76f5af3-9474-4186-9da0-d5ea4fe14f09" containerName="mariadb-account-create" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.667581 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ef1413-475d-4f9b-ac37-24b52ed55ec8" containerName="init" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.668063 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.669902 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p87jt" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.671548 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.680005 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pnqtc"] Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.711420 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snncn\" (UniqueName: \"kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.711554 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.711658 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.813455 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snncn\" (UniqueName: \"kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.813505 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.813543 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.819029 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.821235 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.835664 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snncn\" (UniqueName: \"kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn\") pod \"barbican-db-sync-pnqtc\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.917954 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-2b2jk"] Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.919189 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.923648 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8r9xt" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.923858 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.925305 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.928048 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2b2jk"] Oct 14 07:08:37 crc kubenswrapper[5018]: I1014 07:08:37.987785 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019031 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019086 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019122 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019246 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019330 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.019454 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szr4x\" (UniqueName: \"kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.120749 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.120815 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.120841 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.120855 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.121373 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.121429 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.121485 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szr4x\" (UniqueName: \"kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.126267 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.136639 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.139166 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.139270 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szr4x\" (UniqueName: \"kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.151168 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts\") pod \"cinder-db-sync-2b2jk\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.246727 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.548957 5018 generic.go:334] "Generic (PLEG): container finished" podID="456702b4-2d47-4803-b768-ebc9dcc679eb" containerID="eb63bc0308d56dec44abe934bec14d642df71515d6ac71e9bf8281b994bb49a4" exitCode=0 Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.549032 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r8t2m" event={"ID":"456702b4-2d47-4803-b768-ebc9dcc679eb","Type":"ContainerDied","Data":"eb63bc0308d56dec44abe934bec14d642df71515d6ac71e9bf8281b994bb49a4"} Oct 14 07:08:38 crc kubenswrapper[5018]: I1014 07:08:38.620449 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ef1413-475d-4f9b-ac37-24b52ed55ec8" path="/var/lib/kubelet/pods/88ef1413-475d-4f9b-ac37-24b52ed55ec8/volumes" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.072122 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.175334 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.175808 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.175872 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.175918 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.175950 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.176070 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqzs5\" (UniqueName: \"kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5\") pod \"456702b4-2d47-4803-b768-ebc9dcc679eb\" (UID: \"456702b4-2d47-4803-b768-ebc9dcc679eb\") " Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.183554 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.183585 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.184190 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5" (OuterVolumeSpecName: "kube-api-access-fqzs5") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "kube-api-access-fqzs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.184436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts" (OuterVolumeSpecName: "scripts") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.204762 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data" (OuterVolumeSpecName: "config-data") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.207931 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "456702b4-2d47-4803-b768-ebc9dcc679eb" (UID: "456702b4-2d47-4803-b768-ebc9dcc679eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277906 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqzs5\" (UniqueName: \"kubernetes.io/projected/456702b4-2d47-4803-b768-ebc9dcc679eb-kube-api-access-fqzs5\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277939 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277950 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277959 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277966 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.277977 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/456702b4-2d47-4803-b768-ebc9dcc679eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:41 crc kubenswrapper[5018]: W1014 07:08:41.357800 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25622217_99df_4f81_a099_b17af99a7dd6.slice/crio-213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508 WatchSource:0}: Error finding container 213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508: Status 404 returned error can't find the container with id 213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508 Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.364564 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2b2jk"] Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.439285 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pnqtc"] Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.575534 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2b2jk" event={"ID":"25622217-99df-4f81-a099-b17af99a7dd6","Type":"ContainerStarted","Data":"213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508"} Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.576865 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r8t2m" event={"ID":"456702b4-2d47-4803-b768-ebc9dcc679eb","Type":"ContainerDied","Data":"086044733bdef854a2cd11ab51ecbae5380d66c69118eb362c3e83c01e15eb6a"} Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.576888 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="086044733bdef854a2cd11ab51ecbae5380d66c69118eb362c3e83c01e15eb6a" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.576931 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r8t2m" Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.586453 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pnqtc" event={"ID":"609e45c4-103a-48d3-9113-6335839588b3","Type":"ContainerStarted","Data":"556c52bd2e7654a558bfb4bb85d320f539da752496c30e9eea0678c591220820"} Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.587860 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerStarted","Data":"99b99075e3cc9efca3039a048d40e4309b1ee228733b7b47d0a3b9ce4f8f2730"} Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.589374 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4nzkd" event={"ID":"6920d271-45c6-4d58-af8a-6707e5fa80e9","Type":"ContainerStarted","Data":"75818e8b9032462b8d1fa44368e8a77b40f4d9c29e54cb0b2af0d3c3b89438b7"} Oct 14 07:08:41 crc kubenswrapper[5018]: I1014 07:08:41.613253 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4nzkd" podStartSLOduration=1.750675786 podStartE2EDuration="7.613228828s" podCreationTimestamp="2025-10-14 07:08:34 +0000 UTC" firstStartedPulling="2025-10-14 07:08:35.087054943 +0000 UTC m=+1131.671101570" lastFinishedPulling="2025-10-14 07:08:40.949607975 +0000 UTC m=+1137.533654612" observedRunningTime="2025-10-14 07:08:41.602598456 +0000 UTC m=+1138.186645083" watchObservedRunningTime="2025-10-14 07:08:41.613228828 +0000 UTC m=+1138.197275465" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.220707 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-r8t2m"] Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.229009 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-r8t2m"] Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.342484 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-d7wwp"] Oct 14 07:08:42 crc kubenswrapper[5018]: E1014 07:08:42.343672 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456702b4-2d47-4803-b768-ebc9dcc679eb" containerName="keystone-bootstrap" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.343700 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="456702b4-2d47-4803-b768-ebc9dcc679eb" containerName="keystone-bootstrap" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.343911 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="456702b4-2d47-4803-b768-ebc9dcc679eb" containerName="keystone-bootstrap" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.344545 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.348001 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.348380 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ptvv" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.349107 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.350567 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.378074 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d7wwp"] Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.410552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.410633 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.411108 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.411273 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl5ks\" (UniqueName: \"kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.411920 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.412068 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513168 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl5ks\" (UniqueName: \"kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513461 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513492 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513515 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513531 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.513568 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.519162 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.519890 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.520867 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.521762 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.522195 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.536745 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl5ks\" (UniqueName: \"kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks\") pod \"keystone-bootstrap-d7wwp\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.601682 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerStarted","Data":"d17b1e38a0f52e8c47fd151211bfd4af434143680579ee42c649929b908a86a0"} Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.621206 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456702b4-2d47-4803-b768-ebc9dcc679eb" path="/var/lib/kubelet/pods/456702b4-2d47-4803-b768-ebc9dcc679eb/volumes" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.683908 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-169d-account-create-w72n8"] Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.684847 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.687461 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.704164 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-169d-account-create-w72n8"] Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.716007 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlnqr\" (UniqueName: \"kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr\") pod \"neutron-169d-account-create-w72n8\" (UID: \"c4f7d916-4a85-4465-a5fa-af904b18a50b\") " pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.769470 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.817363 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlnqr\" (UniqueName: \"kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr\") pod \"neutron-169d-account-create-w72n8\" (UID: \"c4f7d916-4a85-4465-a5fa-af904b18a50b\") " pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:42 crc kubenswrapper[5018]: I1014 07:08:42.833146 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlnqr\" (UniqueName: \"kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr\") pod \"neutron-169d-account-create-w72n8\" (UID: \"c4f7d916-4a85-4465-a5fa-af904b18a50b\") " pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.007078 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.196089 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d7wwp"] Oct 14 07:08:43 crc kubenswrapper[5018]: W1014 07:08:43.244229 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f2aa868_3d6a_4ff9_8f40_bb12bf6e4a22.slice/crio-e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73 WatchSource:0}: Error finding container e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73: Status 404 returned error can't find the container with id e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73 Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.429973 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-169d-account-create-w72n8"] Oct 14 07:08:43 crc kubenswrapper[5018]: W1014 07:08:43.432796 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4f7d916_4a85_4465_a5fa_af904b18a50b.slice/crio-4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a WatchSource:0}: Error finding container 4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a: Status 404 returned error can't find the container with id 4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.611445 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-169d-account-create-w72n8" event={"ID":"c4f7d916-4a85-4465-a5fa-af904b18a50b","Type":"ContainerStarted","Data":"9051ceab7493cbb60404bd9f62f8db4f133e15225ee6817015967faf21cef4f7"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.611486 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-169d-account-create-w72n8" event={"ID":"c4f7d916-4a85-4465-a5fa-af904b18a50b","Type":"ContainerStarted","Data":"4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.613510 5018 generic.go:334] "Generic (PLEG): container finished" podID="aad96861-15f8-4715-8b7c-927ae48fe59c" containerID="a76b17433cbbe67cc02ee352389554f84d8ac0be588c5e00e58f7115928d5574" exitCode=0 Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.613662 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-77v4n" event={"ID":"aad96861-15f8-4715-8b7c-927ae48fe59c","Type":"ContainerDied","Data":"a76b17433cbbe67cc02ee352389554f84d8ac0be588c5e00e58f7115928d5574"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.615120 5018 generic.go:334] "Generic (PLEG): container finished" podID="6920d271-45c6-4d58-af8a-6707e5fa80e9" containerID="75818e8b9032462b8d1fa44368e8a77b40f4d9c29e54cb0b2af0d3c3b89438b7" exitCode=0 Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.615199 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4nzkd" event={"ID":"6920d271-45c6-4d58-af8a-6707e5fa80e9","Type":"ContainerDied","Data":"75818e8b9032462b8d1fa44368e8a77b40f4d9c29e54cb0b2af0d3c3b89438b7"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.617906 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d7wwp" event={"ID":"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22","Type":"ContainerStarted","Data":"14bd600b1cdf54ac9e825fe0357c58d8425a49107b37b3119e48f3bee9fb6c6a"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.617947 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d7wwp" event={"ID":"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22","Type":"ContainerStarted","Data":"e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73"} Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.632768 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-169d-account-create-w72n8" podStartSLOduration=1.632746756 podStartE2EDuration="1.632746756s" podCreationTimestamp="2025-10-14 07:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:43.623968125 +0000 UTC m=+1140.208014752" watchObservedRunningTime="2025-10-14 07:08:43.632746756 +0000 UTC m=+1140.216793383" Oct 14 07:08:43 crc kubenswrapper[5018]: I1014 07:08:43.682251 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-d7wwp" podStartSLOduration=1.682234578 podStartE2EDuration="1.682234578s" podCreationTimestamp="2025-10-14 07:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:43.672396798 +0000 UTC m=+1140.256443425" watchObservedRunningTime="2025-10-14 07:08:43.682234578 +0000 UTC m=+1140.266281205" Oct 14 07:08:44 crc kubenswrapper[5018]: I1014 07:08:44.487567 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:08:44 crc kubenswrapper[5018]: I1014 07:08:44.540328 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:44 crc kubenswrapper[5018]: I1014 07:08:44.540551 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" containerID="cri-o://f4e9706118e43b2c9cdcc673466db21989bfa65b18943fa1cbdd21b5134b0bc7" gracePeriod=10 Oct 14 07:08:44 crc kubenswrapper[5018]: I1014 07:08:44.632257 5018 generic.go:334] "Generic (PLEG): container finished" podID="c4f7d916-4a85-4465-a5fa-af904b18a50b" containerID="9051ceab7493cbb60404bd9f62f8db4f133e15225ee6817015967faf21cef4f7" exitCode=0 Oct 14 07:08:44 crc kubenswrapper[5018]: I1014 07:08:44.632480 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-169d-account-create-w72n8" event={"ID":"c4f7d916-4a85-4465-a5fa-af904b18a50b","Type":"ContainerDied","Data":"9051ceab7493cbb60404bd9f62f8db4f133e15225ee6817015967faf21cef4f7"} Oct 14 07:08:45 crc kubenswrapper[5018]: I1014 07:08:45.646601 5018 generic.go:334] "Generic (PLEG): container finished" podID="5b17231c-1edb-4639-8333-678e3434f4dd" containerID="f4e9706118e43b2c9cdcc673466db21989bfa65b18943fa1cbdd21b5134b0bc7" exitCode=0 Oct 14 07:08:45 crc kubenswrapper[5018]: I1014 07:08:45.646780 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" event={"ID":"5b17231c-1edb-4639-8333-678e3434f4dd","Type":"ContainerDied","Data":"f4e9706118e43b2c9cdcc673466db21989bfa65b18943fa1cbdd21b5134b0bc7"} Oct 14 07:08:46 crc kubenswrapper[5018]: I1014 07:08:46.657592 5018 generic.go:334] "Generic (PLEG): container finished" podID="8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" containerID="14bd600b1cdf54ac9e825fe0357c58d8425a49107b37b3119e48f3bee9fb6c6a" exitCode=0 Oct 14 07:08:46 crc kubenswrapper[5018]: I1014 07:08:46.657675 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d7wwp" event={"ID":"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22","Type":"ContainerDied","Data":"14bd600b1cdf54ac9e825fe0357c58d8425a49107b37b3119e48f3bee9fb6c6a"} Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.621297 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.643055 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.650885 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.694420 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4nzkd" event={"ID":"6920d271-45c6-4d58-af8a-6707e5fa80e9","Type":"ContainerDied","Data":"ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332"} Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.694486 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5f82ce833a237bcf919f5bdd06583e811212ef3f13fbba5f1a69d9c4b65332" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.694450 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4nzkd" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.696393 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-169d-account-create-w72n8" event={"ID":"c4f7d916-4a85-4465-a5fa-af904b18a50b","Type":"ContainerDied","Data":"4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a"} Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.696455 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e562a5004335125866524a9fb7e8cd4d40cb7edd7d5a1223344553d925d850a" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.696577 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-169d-account-create-w72n8" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.698340 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-77v4n" event={"ID":"aad96861-15f8-4715-8b7c-927ae48fe59c","Type":"ContainerDied","Data":"2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a"} Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.698347 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-77v4n" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.698362 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2379a0c8403472ec30e0d617d2eb1bd79477a3f3440f6de3c0e0d1e4a0611f7a" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.719044 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlnqr\" (UniqueName: \"kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr\") pod \"c4f7d916-4a85-4465-a5fa-af904b18a50b\" (UID: \"c4f7d916-4a85-4465-a5fa-af904b18a50b\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.749794 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr" (OuterVolumeSpecName: "kube-api-access-xlnqr") pod "c4f7d916-4a85-4465-a5fa-af904b18a50b" (UID: "c4f7d916-4a85-4465-a5fa-af904b18a50b"). InnerVolumeSpecName "kube-api-access-xlnqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.820746 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle\") pod \"aad96861-15f8-4715-8b7c-927ae48fe59c\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821196 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxc99\" (UniqueName: \"kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99\") pod \"6920d271-45c6-4d58-af8a-6707e5fa80e9\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821429 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data\") pod \"6920d271-45c6-4d58-af8a-6707e5fa80e9\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821849 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxkxd\" (UniqueName: \"kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd\") pod \"aad96861-15f8-4715-8b7c-927ae48fe59c\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821916 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs\") pod \"6920d271-45c6-4d58-af8a-6707e5fa80e9\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821982 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts\") pod \"6920d271-45c6-4d58-af8a-6707e5fa80e9\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.821999 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle\") pod \"6920d271-45c6-4d58-af8a-6707e5fa80e9\" (UID: \"6920d271-45c6-4d58-af8a-6707e5fa80e9\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.822264 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs" (OuterVolumeSpecName: "logs") pod "6920d271-45c6-4d58-af8a-6707e5fa80e9" (UID: "6920d271-45c6-4d58-af8a-6707e5fa80e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.822022 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data\") pod \"aad96861-15f8-4715-8b7c-927ae48fe59c\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.822669 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data\") pod \"aad96861-15f8-4715-8b7c-927ae48fe59c\" (UID: \"aad96861-15f8-4715-8b7c-927ae48fe59c\") " Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.823130 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6920d271-45c6-4d58-af8a-6707e5fa80e9-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.823149 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlnqr\" (UniqueName: \"kubernetes.io/projected/c4f7d916-4a85-4465-a5fa-af904b18a50b-kube-api-access-xlnqr\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.826735 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd" (OuterVolumeSpecName: "kube-api-access-fxkxd") pod "aad96861-15f8-4715-8b7c-927ae48fe59c" (UID: "aad96861-15f8-4715-8b7c-927ae48fe59c"). InnerVolumeSpecName "kube-api-access-fxkxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.831741 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts" (OuterVolumeSpecName: "scripts") pod "6920d271-45c6-4d58-af8a-6707e5fa80e9" (UID: "6920d271-45c6-4d58-af8a-6707e5fa80e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.842936 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "aad96861-15f8-4715-8b7c-927ae48fe59c" (UID: "aad96861-15f8-4715-8b7c-927ae48fe59c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.843561 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99" (OuterVolumeSpecName: "kube-api-access-xxc99") pod "6920d271-45c6-4d58-af8a-6707e5fa80e9" (UID: "6920d271-45c6-4d58-af8a-6707e5fa80e9"). InnerVolumeSpecName "kube-api-access-xxc99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.847418 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6920d271-45c6-4d58-af8a-6707e5fa80e9" (UID: "6920d271-45c6-4d58-af8a-6707e5fa80e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.865850 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aad96861-15f8-4715-8b7c-927ae48fe59c" (UID: "aad96861-15f8-4715-8b7c-927ae48fe59c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.874516 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data" (OuterVolumeSpecName: "config-data") pod "6920d271-45c6-4d58-af8a-6707e5fa80e9" (UID: "6920d271-45c6-4d58-af8a-6707e5fa80e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.891400 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data" (OuterVolumeSpecName: "config-data") pod "aad96861-15f8-4715-8b7c-927ae48fe59c" (UID: "aad96861-15f8-4715-8b7c-927ae48fe59c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924510 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924543 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924556 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924571 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924584 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad96861-15f8-4715-8b7c-927ae48fe59c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924597 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxc99\" (UniqueName: \"kubernetes.io/projected/6920d271-45c6-4d58-af8a-6707e5fa80e9-kube-api-access-xxc99\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924609 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6920d271-45c6-4d58-af8a-6707e5fa80e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:47 crc kubenswrapper[5018]: I1014 07:08:47.924637 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxkxd\" (UniqueName: \"kubernetes.io/projected/aad96861-15f8-4715-8b7c-927ae48fe59c-kube-api-access-fxkxd\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.744558 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:08:48 crc kubenswrapper[5018]: E1014 07:08:48.745305 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f7d916-4a85-4465-a5fa-af904b18a50b" containerName="mariadb-account-create" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.745322 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f7d916-4a85-4465-a5fa-af904b18a50b" containerName="mariadb-account-create" Oct 14 07:08:48 crc kubenswrapper[5018]: E1014 07:08:48.745345 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" containerName="glance-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.745353 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" containerName="glance-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: E1014 07:08:48.745377 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6920d271-45c6-4d58-af8a-6707e5fa80e9" containerName="placement-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.745387 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6920d271-45c6-4d58-af8a-6707e5fa80e9" containerName="placement-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.746704 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6920d271-45c6-4d58-af8a-6707e5fa80e9" containerName="placement-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.746740 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f7d916-4a85-4465-a5fa-af904b18a50b" containerName="mariadb-account-create" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.746755 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" containerName="glance-db-sync" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.747930 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.755213 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.755462 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.755588 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.755801 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vpkpb" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.756141 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.761384 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.857673 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.857958 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.857986 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pkhw\" (UniqueName: \"kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.858012 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.858030 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.858083 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.858115 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.960230 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.960289 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.960364 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.961518 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pkhw\" (UniqueName: \"kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.961588 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.961640 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.962891 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.961756 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.971561 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.972097 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.972530 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.973406 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.984340 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:48 crc kubenswrapper[5018]: I1014 07:08:48.984588 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pkhw\" (UniqueName: \"kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw\") pod \"placement-7649bf8b6b-5szh6\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.014998 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.025431 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.029703 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.127663 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170291 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170342 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170371 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170531 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpjdr\" (UniqueName: \"kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170713 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.170741 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.271878 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.271939 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.271964 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.271999 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpjdr\" (UniqueName: \"kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.272020 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.272041 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.272876 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.273388 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.273863 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.274191 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.274428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.303777 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpjdr\" (UniqueName: \"kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr\") pod \"dnsmasq-dns-77f7885f7f-f7d66\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:49 crc kubenswrapper[5018]: I1014 07:08:49.357726 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.079426 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.080866 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.082364 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.082777 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.083894 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pj22j" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.090223 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.131728 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.133110 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.136741 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.141746 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.187863 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.187901 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fmdn\" (UniqueName: \"kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.187939 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.187984 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.188005 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.188074 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.188094 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.289914 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290010 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290035 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fmdn\" (UniqueName: \"kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290084 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290126 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290156 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290180 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290212 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290271 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290311 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290337 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290389 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.290436 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jf4w\" (UniqueName: \"kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.292174 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.294241 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.294997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.299244 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.313803 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fmdn\" (UniqueName: \"kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.328383 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.331941 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.359124 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392185 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jf4w\" (UniqueName: \"kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392528 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392589 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392612 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392647 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392679 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392718 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.392857 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.393506 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.393506 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.397114 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.398576 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.413254 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.414485 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.417396 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.421696 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jf4w\" (UniqueName: \"kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w\") pod \"glance-default-internal-api-0\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:08:50 crc kubenswrapper[5018]: I1014 07:08:50.447631 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:08:51 crc kubenswrapper[5018]: I1014 07:08:51.285183 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:08:51 crc kubenswrapper[5018]: I1014 07:08:51.357214 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.897654 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nx7zt"] Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.899064 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.907217 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nx7zt"] Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.908113 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-k7ccr" Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.908365 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 07:08:52 crc kubenswrapper[5018]: I1014 07:08:52.908599 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.041919 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhvbc\" (UniqueName: \"kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.042024 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.042058 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.143927 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.143974 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.144094 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhvbc\" (UniqueName: \"kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.150189 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.150449 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.175824 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhvbc\" (UniqueName: \"kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc\") pod \"neutron-db-sync-nx7zt\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.231142 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:08:53 crc kubenswrapper[5018]: I1014 07:08:53.681032 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: i/o timeout" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.544291 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.550028 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642527 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642595 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nppb5\" (UniqueName: \"kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642641 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl5ks\" (UniqueName: \"kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642714 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642751 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642816 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642885 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642913 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642937 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.642992 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.643015 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc\") pod \"5b17231c-1edb-4639-8333-678e3434f4dd\" (UID: \"5b17231c-1edb-4639-8333-678e3434f4dd\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.643041 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle\") pod \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\" (UID: \"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22\") " Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.651070 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.651366 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks" (OuterVolumeSpecName: "kube-api-access-hl5ks") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "kube-api-access-hl5ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.651509 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5" (OuterVolumeSpecName: "kube-api-access-nppb5") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "kube-api-access-nppb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.655794 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.656209 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts" (OuterVolumeSpecName: "scripts") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.684993 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data" (OuterVolumeSpecName: "config-data") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.685204 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" (UID: "8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.708720 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.714894 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.716206 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config" (OuterVolumeSpecName: "config") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.717088 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.719870 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b17231c-1edb-4639-8333-678e3434f4dd" (UID: "5b17231c-1edb-4639-8333-678e3434f4dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745585 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745651 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nppb5\" (UniqueName: \"kubernetes.io/projected/5b17231c-1edb-4639-8333-678e3434f4dd-kube-api-access-nppb5\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745669 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl5ks\" (UniqueName: \"kubernetes.io/projected/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-kube-api-access-hl5ks\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745686 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745700 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745780 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745807 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745816 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745828 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745839 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745847 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b17231c-1edb-4639-8333-678e3434f4dd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.745856 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.797091 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d7wwp" event={"ID":"8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22","Type":"ContainerDied","Data":"e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73"} Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.797124 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e41764ffba15b38963fc0f26d401be645a36907cbb3bd83e42d97a3b8b201b73" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.797138 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d7wwp" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.806230 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" event={"ID":"5b17231c-1edb-4639-8333-678e3434f4dd","Type":"ContainerDied","Data":"bf7603f5e713fba4b1f9dd355a1af92e7fd57dba43286967a1fe9629fc4abfd6"} Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.806293 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.806302 5018 scope.go:117] "RemoveContainer" containerID="f4e9706118e43b2c9cdcc673466db21989bfa65b18943fa1cbdd21b5134b0bc7" Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.850214 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:57 crc kubenswrapper[5018]: I1014 07:08:57.859677 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d9ddcb47c-qm2ls"] Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.617230 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" path="/var/lib/kubelet/pods/5b17231c-1edb-4639-8333-678e3434f4dd/volumes" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.672818 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.673229 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="init" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.673249 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="init" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.673267 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" containerName="keystone-bootstrap" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.673276 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" containerName="keystone-bootstrap" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.673309 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.673320 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.673549 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.673593 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" containerName="keystone-bootstrap" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.674241 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680117 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680129 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680461 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680461 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680533 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6ptvv" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.680708 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.681493 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-d9ddcb47c-qm2ls" podUID="5b17231c-1edb-4639-8333-678e3434f4dd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: i/o timeout" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.695563 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.762680 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.762975 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml26k\" (UniqueName: \"kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763296 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763412 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763510 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763608 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763752 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.763920 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.782980 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.783137 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-szr4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-2b2jk_openstack(25622217-99df-4f81-a099-b17af99a7dd6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.784370 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-2b2jk" podUID="25622217-99df-4f81-a099-b17af99a7dd6" Oct 14 07:08:58 crc kubenswrapper[5018]: E1014 07:08:58.821595 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429\\\"\"" pod="openstack/cinder-db-sync-2b2jk" podUID="25622217-99df-4f81-a099-b17af99a7dd6" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866122 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866184 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866199 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866216 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866255 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866311 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866391 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.866428 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml26k\" (UniqueName: \"kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.871385 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.871821 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.872088 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.872244 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.872708 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.872990 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.873690 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.883511 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml26k\" (UniqueName: \"kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k\") pod \"keystone-5ff4499-6wpcq\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:08:58 crc kubenswrapper[5018]: I1014 07:08:58.991602 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:09:00 crc kubenswrapper[5018]: E1014 07:09:00.228878 5018 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1" Oct 14 07:09:00 crc kubenswrapper[5018]: E1014 07:09:00.229019 5018 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s8rsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(2af870e0-db4f-49ba-afeb-238282c56c61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.273846 5018 scope.go:117] "RemoveContainer" containerID="60a1142fe263841ea85847e103545520b46aee8733cb982cb5e8bd7b8fc70379" Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.835811 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.841820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pnqtc" event={"ID":"609e45c4-103a-48d3-9113-6335839588b3","Type":"ContainerStarted","Data":"c60820610e6717de5af409b5508e7a1c494ce74fdb1fcce6f6fbbec19e7ed673"} Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.879424 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pnqtc" podStartSLOduration=5.08690993 podStartE2EDuration="23.879399647s" podCreationTimestamp="2025-10-14 07:08:37 +0000 UTC" firstStartedPulling="2025-10-14 07:08:41.437406659 +0000 UTC m=+1138.021453286" lastFinishedPulling="2025-10-14 07:09:00.229896376 +0000 UTC m=+1156.813943003" observedRunningTime="2025-10-14 07:09:00.863733743 +0000 UTC m=+1157.447780370" watchObservedRunningTime="2025-10-14 07:09:00.879399647 +0000 UTC m=+1157.463446274" Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.899009 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:09:00 crc kubenswrapper[5018]: W1014 07:09:00.904256 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53363d98_901b_4334_bfcc_a74638255459.slice/crio-6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4 WatchSource:0}: Error finding container 6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4: Status 404 returned error can't find the container with id 6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4 Oct 14 07:09:00 crc kubenswrapper[5018]: W1014 07:09:00.904762 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5711950b_f891_4fc2_84fe_fd0c12d1b228.slice/crio-154a46a7632e7db0bc67efc984db7da10180b937c5f6943d75c27466a3372732 WatchSource:0}: Error finding container 154a46a7632e7db0bc67efc984db7da10180b937c5f6943d75c27466a3372732: Status 404 returned error can't find the container with id 154a46a7632e7db0bc67efc984db7da10180b937c5f6943d75c27466a3372732 Oct 14 07:09:00 crc kubenswrapper[5018]: I1014 07:09:00.905324 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nx7zt"] Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.011010 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:09:01 crc kubenswrapper[5018]: W1014 07:09:01.016768 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod390ed4c4_a917_47a0_83e5_5c05fbaf2005.slice/crio-5aa41398fc676eceb74ac034937b5650afc9d535a7d583a4d5d68e3dd17ce981 WatchSource:0}: Error finding container 5aa41398fc676eceb74ac034937b5650afc9d535a7d583a4d5d68e3dd17ce981: Status 404 returned error can't find the container with id 5aa41398fc676eceb74ac034937b5650afc9d535a7d583a4d5d68e3dd17ce981 Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.030821 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:01 crc kubenswrapper[5018]: W1014 07:09:01.046753 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ecd3ea9_98af_4b53_a40a_8b5b1c48da41.slice/crio-f655bf0af755c40d94be867036fee4298aaecf84f2d7d8846d93d937db52ef78 WatchSource:0}: Error finding container f655bf0af755c40d94be867036fee4298aaecf84f2d7d8846d93d937db52ef78: Status 404 returned error can't find the container with id f655bf0af755c40d94be867036fee4298aaecf84f2d7d8846d93d937db52ef78 Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.127988 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:01 crc kubenswrapper[5018]: W1014 07:09:01.142896 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44751d49_a17c_4e9e_b1df_e2710d5dc481.slice/crio-39c54dc1ec1d37fe578c9bc778f2204d662ec6d5fc27279d82b39018a58a25d4 WatchSource:0}: Error finding container 39c54dc1ec1d37fe578c9bc778f2204d662ec6d5fc27279d82b39018a58a25d4: Status 404 returned error can't find the container with id 39c54dc1ec1d37fe578c9bc778f2204d662ec6d5fc27279d82b39018a58a25d4 Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.854409 5018 generic.go:334] "Generic (PLEG): container finished" podID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerID="f3f27d2120e9963fe5aec28385b0c7f1624e4a6ec9ed9738212db3391674a203" exitCode=0 Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.854721 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" event={"ID":"5711950b-f891-4fc2-84fe-fd0c12d1b228","Type":"ContainerDied","Data":"f3f27d2120e9963fe5aec28385b0c7f1624e4a6ec9ed9738212db3391674a203"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.854750 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" event={"ID":"5711950b-f891-4fc2-84fe-fd0c12d1b228","Type":"ContainerStarted","Data":"154a46a7632e7db0bc67efc984db7da10180b937c5f6943d75c27466a3372732"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.858142 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerStarted","Data":"bfd96cd797a9cef67b6577318e35f224798021a3e38496f28eb79e625754ac78"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.858178 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerStarted","Data":"f655bf0af755c40d94be867036fee4298aaecf84f2d7d8846d93d937db52ef78"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.860279 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nx7zt" event={"ID":"53363d98-901b-4334-bfcc-a74638255459","Type":"ContainerStarted","Data":"a69669374cda679bf77ce540f42dd59f2d251b3623db67fb8c294ce9d28fd5b8"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.860310 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nx7zt" event={"ID":"53363d98-901b-4334-bfcc-a74638255459","Type":"ContainerStarted","Data":"6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.862756 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerStarted","Data":"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.862800 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerStarted","Data":"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.862811 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerStarted","Data":"fcc6bfc97a2b39a7c251e57a72e8f5d749619f8452b5d6864d9ec92f5c3efbd8"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.862850 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.863316 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.865048 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerStarted","Data":"b9b31d28ad5ef69f136cd455ee32490afde1ccd6e3aea1ed059cbe782593af82"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.865090 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerStarted","Data":"39c54dc1ec1d37fe578c9bc778f2204d662ec6d5fc27279d82b39018a58a25d4"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.868144 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ff4499-6wpcq" event={"ID":"390ed4c4-a917-47a0-83e5-5c05fbaf2005","Type":"ContainerStarted","Data":"b483c2ab5ee4494a8c3c6374f47798895d55ff335e8b914c8d3d4714c801eb88"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.868277 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ff4499-6wpcq" event={"ID":"390ed4c4-a917-47a0-83e5-5c05fbaf2005","Type":"ContainerStarted","Data":"5aa41398fc676eceb74ac034937b5650afc9d535a7d583a4d5d68e3dd17ce981"} Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.869079 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.892642 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nx7zt" podStartSLOduration=9.892608702 podStartE2EDuration="9.892608702s" podCreationTimestamp="2025-10-14 07:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:01.887803316 +0000 UTC m=+1158.471849943" watchObservedRunningTime="2025-10-14 07:09:01.892608702 +0000 UTC m=+1158.476655329" Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.914106 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7649bf8b6b-5szh6" podStartSLOduration=13.91408995 podStartE2EDuration="13.91408995s" podCreationTimestamp="2025-10-14 07:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:01.903821729 +0000 UTC m=+1158.487868356" watchObservedRunningTime="2025-10-14 07:09:01.91408995 +0000 UTC m=+1158.498136567" Oct 14 07:09:01 crc kubenswrapper[5018]: I1014 07:09:01.932769 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5ff4499-6wpcq" podStartSLOduration=3.932753989 podStartE2EDuration="3.932753989s" podCreationTimestamp="2025-10-14 07:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:01.929779155 +0000 UTC m=+1158.513825782" watchObservedRunningTime="2025-10-14 07:09:01.932753989 +0000 UTC m=+1158.516800606" Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.464080 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.464126 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.464164 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.464706 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.464753 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd" gracePeriod=600 Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.890667 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerStarted","Data":"7f72f7ca9b80406586a87e9b3fb1884ecc1bbbb91e73838acf361aaa64307299"} Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.894497 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd" exitCode=0 Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.894552 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd"} Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.894578 5018 scope.go:117] "RemoveContainer" containerID="51067037abdc4720de85c8704cb1f500ab56bc97bff15c02e74f6d36edb40986" Oct 14 07:09:02 crc kubenswrapper[5018]: I1014 07:09:02.901366 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" event={"ID":"5711950b-f891-4fc2-84fe-fd0c12d1b228","Type":"ContainerStarted","Data":"d6662481997568702efbfc8fa1f8bfc82f15457b8a4d5500c6fb2edaf4f13b56"} Oct 14 07:09:03 crc kubenswrapper[5018]: E1014 07:09:03.043913 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.910134 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerStarted","Data":"777701be4a1b35e6222e836b636046f71af77efe74c3e630f85c8067ea4696b4"} Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.910902 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-log" containerID="cri-o://b9b31d28ad5ef69f136cd455ee32490afde1ccd6e3aea1ed059cbe782593af82" gracePeriod=30 Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.911297 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-httpd" containerID="cri-o://777701be4a1b35e6222e836b636046f71af77efe74c3e630f85c8067ea4696b4" gracePeriod=30 Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.913882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d"} Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.913928 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-log" containerID="cri-o://bfd96cd797a9cef67b6577318e35f224798021a3e38496f28eb79e625754ac78" gracePeriod=30 Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.913997 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-httpd" containerID="cri-o://7f72f7ca9b80406586a87e9b3fb1884ecc1bbbb91e73838acf361aaa64307299" gracePeriod=30 Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.914272 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.941065 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=14.941047124 podStartE2EDuration="14.941047124s" podCreationTimestamp="2025-10-14 07:08:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:03.937696689 +0000 UTC m=+1160.521743316" watchObservedRunningTime="2025-10-14 07:09:03.941047124 +0000 UTC m=+1160.525093751" Oct 14 07:09:03 crc kubenswrapper[5018]: I1014 07:09:03.967335 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=14.967318947999999 podStartE2EDuration="14.967318948s" podCreationTimestamp="2025-10-14 07:08:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:03.966754262 +0000 UTC m=+1160.550800909" watchObservedRunningTime="2025-10-14 07:09:03.967318948 +0000 UTC m=+1160.551365565" Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.004393 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" podStartSLOduration=16.004376548 podStartE2EDuration="16.004376548s" podCreationTimestamp="2025-10-14 07:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:04.001082005 +0000 UTC m=+1160.585128642" watchObservedRunningTime="2025-10-14 07:09:04.004376548 +0000 UTC m=+1160.588423175" Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.925888 5018 generic.go:334] "Generic (PLEG): container finished" podID="609e45c4-103a-48d3-9113-6335839588b3" containerID="c60820610e6717de5af409b5508e7a1c494ce74fdb1fcce6f6fbbec19e7ed673" exitCode=0 Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.926230 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pnqtc" event={"ID":"609e45c4-103a-48d3-9113-6335839588b3","Type":"ContainerDied","Data":"c60820610e6717de5af409b5508e7a1c494ce74fdb1fcce6f6fbbec19e7ed673"} Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.930847 5018 generic.go:334] "Generic (PLEG): container finished" podID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerID="7f72f7ca9b80406586a87e9b3fb1884ecc1bbbb91e73838acf361aaa64307299" exitCode=0 Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.930879 5018 generic.go:334] "Generic (PLEG): container finished" podID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerID="bfd96cd797a9cef67b6577318e35f224798021a3e38496f28eb79e625754ac78" exitCode=143 Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.930924 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerDied","Data":"7f72f7ca9b80406586a87e9b3fb1884ecc1bbbb91e73838acf361aaa64307299"} Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.930951 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerDied","Data":"bfd96cd797a9cef67b6577318e35f224798021a3e38496f28eb79e625754ac78"} Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.933179 5018 generic.go:334] "Generic (PLEG): container finished" podID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerID="777701be4a1b35e6222e836b636046f71af77efe74c3e630f85c8067ea4696b4" exitCode=0 Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.933202 5018 generic.go:334] "Generic (PLEG): container finished" podID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerID="b9b31d28ad5ef69f136cd455ee32490afde1ccd6e3aea1ed059cbe782593af82" exitCode=143 Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.934415 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerDied","Data":"777701be4a1b35e6222e836b636046f71af77efe74c3e630f85c8067ea4696b4"} Oct 14 07:09:04 crc kubenswrapper[5018]: I1014 07:09:04.934448 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerDied","Data":"b9b31d28ad5ef69f136cd455ee32490afde1ccd6e3aea1ed059cbe782593af82"} Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.269811 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.368672 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snncn\" (UniqueName: \"kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn\") pod \"609e45c4-103a-48d3-9113-6335839588b3\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.368797 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data\") pod \"609e45c4-103a-48d3-9113-6335839588b3\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.368873 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle\") pod \"609e45c4-103a-48d3-9113-6335839588b3\" (UID: \"609e45c4-103a-48d3-9113-6335839588b3\") " Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.375960 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn" (OuterVolumeSpecName: "kube-api-access-snncn") pod "609e45c4-103a-48d3-9113-6335839588b3" (UID: "609e45c4-103a-48d3-9113-6335839588b3"). InnerVolumeSpecName "kube-api-access-snncn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.376196 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "609e45c4-103a-48d3-9113-6335839588b3" (UID: "609e45c4-103a-48d3-9113-6335839588b3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.405738 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "609e45c4-103a-48d3-9113-6335839588b3" (UID: "609e45c4-103a-48d3-9113-6335839588b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.470760 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.470792 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609e45c4-103a-48d3-9113-6335839588b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.470802 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snncn\" (UniqueName: \"kubernetes.io/projected/609e45c4-103a-48d3-9113-6335839588b3-kube-api-access-snncn\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.966051 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pnqtc" event={"ID":"609e45c4-103a-48d3-9113-6335839588b3","Type":"ContainerDied","Data":"556c52bd2e7654a558bfb4bb85d320f539da752496c30e9eea0678c591220820"} Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.966100 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pnqtc" Oct 14 07:09:07 crc kubenswrapper[5018]: I1014 07:09:07.966108 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="556c52bd2e7654a558bfb4bb85d320f539da752496c30e9eea0678c591220820" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.560356 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:09:08 crc kubenswrapper[5018]: E1014 07:09:08.561110 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609e45c4-103a-48d3-9113-6335839588b3" containerName="barbican-db-sync" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.561128 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="609e45c4-103a-48d3-9113-6335839588b3" containerName="barbican-db-sync" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.561375 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="609e45c4-103a-48d3-9113-6335839588b3" containerName="barbican-db-sync" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.562526 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.573509 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.573687 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.573698 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-p87jt" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.592022 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.645662 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.645877 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="dnsmasq-dns" containerID="cri-o://d6662481997568702efbfc8fa1f8bfc82f15457b8a4d5500c6fb2edaf4f13b56" gracePeriod=10 Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.646804 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.655635 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.657045 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.667322 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.668114 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.692824 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqsbx\" (UniqueName: \"kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.692885 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.692917 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.692993 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.693016 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.708747 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.710329 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.727920 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797197 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797256 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqsbx\" (UniqueName: \"kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797280 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797302 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhk5w\" (UniqueName: \"kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797323 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797359 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797383 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797402 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797423 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797442 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797465 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797500 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797514 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchgt\" (UniqueName: \"kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797555 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797579 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.797597 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.798324 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.804293 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.804846 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.818529 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.825983 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.830067 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqsbx\" (UniqueName: \"kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx\") pod \"barbican-worker-6d56c684b7-gtf99\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.835073 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.839637 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.906390 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913527 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913689 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913761 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913807 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhk5w\" (UniqueName: \"kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913834 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913890 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913917 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913958 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.913998 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.914017 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.914071 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.914092 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchgt\" (UniqueName: \"kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.915369 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.915370 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.916603 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.916995 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.917469 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.918120 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.930071 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.931006 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.939219 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.950476 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhk5w\" (UniqueName: \"kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w\") pod \"barbican-keystone-listener-5d9d59b97b-ng574\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.967854 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchgt\" (UniqueName: \"kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt\") pod \"dnsmasq-dns-5455955775-hs6f6\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:08 crc kubenswrapper[5018]: I1014 07:09:08.983581 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.018563 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.018645 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.018678 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d77vf\" (UniqueName: \"kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.018697 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.018732 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.033844 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.036833 5018 generic.go:334] "Generic (PLEG): container finished" podID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerID="d6662481997568702efbfc8fa1f8bfc82f15457b8a4d5500c6fb2edaf4f13b56" exitCode=0 Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.036869 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" event={"ID":"5711950b-f891-4fc2-84fe-fd0c12d1b228","Type":"ContainerDied","Data":"d6662481997568702efbfc8fa1f8bfc82f15457b8a4d5500c6fb2edaf4f13b56"} Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120192 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d77vf\" (UniqueName: \"kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120234 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120279 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120405 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120559 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.120829 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.126913 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.127539 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.128365 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.138809 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d77vf\" (UniqueName: \"kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf\") pod \"barbican-api-7bd6845fdb-hbp48\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.151040 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.153250 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.205316 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324587 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324660 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324689 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324761 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324782 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324831 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324857 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324873 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324894 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jf4w\" (UniqueName: \"kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324927 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324943 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324965 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fmdn\" (UniqueName: \"kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.324986 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle\") pod \"44751d49-a17c-4e9e-b1df-e2710d5dc481\" (UID: \"44751d49-a17c-4e9e-b1df-e2710d5dc481\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.325044 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs\") pod \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\" (UID: \"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.325586 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs" (OuterVolumeSpecName: "logs") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.325740 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs" (OuterVolumeSpecName: "logs") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.326424 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.330119 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.330936 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts" (OuterVolumeSpecName: "scripts") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.335003 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.339431 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn" (OuterVolumeSpecName: "kube-api-access-4fmdn") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "kube-api-access-4fmdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.341613 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts" (OuterVolumeSpecName: "scripts") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.341841 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w" (OuterVolumeSpecName: "kube-api-access-7jf4w") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "kube-api-access-7jf4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.349099 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.352534 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:09:09 crc kubenswrapper[5018]: E1014 07:09:09.402404 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432740 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432773 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432786 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432797 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jf4w\" (UniqueName: \"kubernetes.io/projected/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-kube-api-access-7jf4w\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432807 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44751d49-a17c-4e9e-b1df-e2710d5dc481-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432814 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432823 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fmdn\" (UniqueName: \"kubernetes.io/projected/44751d49-a17c-4e9e-b1df-e2710d5dc481-kube-api-access-4fmdn\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432831 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432838 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.432846 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.437738 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.457045 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.461790 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data" (OuterVolumeSpecName: "config-data") pod "44751d49-a17c-4e9e-b1df-e2710d5dc481" (UID: "44751d49-a17c-4e9e-b1df-e2710d5dc481"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.470739 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.485584 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data" (OuterVolumeSpecName: "config-data") pod "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" (UID: "0ecd3ea9-98af-4b53-a40a-8b5b1c48da41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.501781 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534353 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534472 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534513 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534573 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534659 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.534707 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpjdr\" (UniqueName: \"kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr\") pod \"5711950b-f891-4fc2-84fe-fd0c12d1b228\" (UID: \"5711950b-f891-4fc2-84fe-fd0c12d1b228\") " Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535110 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535129 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535139 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535148 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535158 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.535166 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44751d49-a17c-4e9e-b1df-e2710d5dc481-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.540412 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr" (OuterVolumeSpecName: "kube-api-access-hpjdr") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "kube-api-access-hpjdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.579673 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config" (OuterVolumeSpecName: "config") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.582718 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.583634 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.588080 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.589569 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5711950b-f891-4fc2-84fe-fd0c12d1b228" (UID: "5711950b-f891-4fc2-84fe-fd0c12d1b228"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637669 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpjdr\" (UniqueName: \"kubernetes.io/projected/5711950b-f891-4fc2-84fe-fd0c12d1b228-kube-api-access-hpjdr\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637712 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637730 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637748 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637765 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.637783 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5711950b-f891-4fc2-84fe-fd0c12d1b228-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.697112 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:09:09 crc kubenswrapper[5018]: W1014 07:09:09.857781 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda49d558a_c4f6_49d3_8210_c38799ab89c4.slice/crio-e00074b1a0eed139d4d4c3237a8c52c563bac8cdbc8c7530d7301cf8141f8925 WatchSource:0}: Error finding container e00074b1a0eed139d4d4c3237a8c52c563bac8cdbc8c7530d7301cf8141f8925: Status 404 returned error can't find the container with id e00074b1a0eed139d4d4c3237a8c52c563bac8cdbc8c7530d7301cf8141f8925 Oct 14 07:09:09 crc kubenswrapper[5018]: W1014 07:09:09.859870 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e54cbba_6800_47af_8d42_fcb95bcf621e.slice/crio-8e105987145d656de6eed8c37a3020c3937a4492e628cec57be189dd9636dc2c WatchSource:0}: Error finding container 8e105987145d656de6eed8c37a3020c3937a4492e628cec57be189dd9636dc2c: Status 404 returned error can't find the container with id 8e105987145d656de6eed8c37a3020c3937a4492e628cec57be189dd9636dc2c Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.879695 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.890053 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:09:09 crc kubenswrapper[5018]: I1014 07:09:09.897286 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.046608 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ecd3ea9-98af-4b53-a40a-8b5b1c48da41","Type":"ContainerDied","Data":"f655bf0af755c40d94be867036fee4298aaecf84f2d7d8846d93d937db52ef78"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.046712 5018 scope.go:117] "RemoveContainer" containerID="7f72f7ca9b80406586a87e9b3fb1884ecc1bbbb91e73838acf361aaa64307299" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.046645 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.047717 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerStarted","Data":"e00074b1a0eed139d4d4c3237a8c52c563bac8cdbc8c7530d7301cf8141f8925"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.049741 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerStarted","Data":"20b6b92a2174a2ef0fe0c4f5b0cf89dc09edc6f8f9a30dd835bb36436c9d3209"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.052939 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.052955 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"44751d49-a17c-4e9e-b1df-e2710d5dc481","Type":"ContainerDied","Data":"39c54dc1ec1d37fe578c9bc778f2204d662ec6d5fc27279d82b39018a58a25d4"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.055208 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerStarted","Data":"3d67473c73ce5ff0b49ff7efeeb23d11ea90445d253f7e91ae84e9887ff6527e"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.058193 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5455955775-hs6f6" event={"ID":"7e54cbba-6800-47af-8d42-fcb95bcf621e","Type":"ContainerStarted","Data":"8e105987145d656de6eed8c37a3020c3937a4492e628cec57be189dd9636dc2c"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.060472 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerStarted","Data":"c32866d679f7b31297ab020b785b7625fa428a5926da8f4decfc634fc65e0844"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.060681 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-central-agent" containerID="cri-o://99b99075e3cc9efca3039a048d40e4309b1ee228733b7b47d0a3b9ce4f8f2730" gracePeriod=30 Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.060988 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.061294 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="proxy-httpd" containerID="cri-o://c32866d679f7b31297ab020b785b7625fa428a5926da8f4decfc634fc65e0844" gracePeriod=30 Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.061356 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-notification-agent" containerID="cri-o://d17b1e38a0f52e8c47fd151211bfd4af434143680579ee42c649929b908a86a0" gracePeriod=30 Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.070993 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" event={"ID":"5711950b-f891-4fc2-84fe-fd0c12d1b228","Type":"ContainerDied","Data":"154a46a7632e7db0bc67efc984db7da10180b937c5f6943d75c27466a3372732"} Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.071039 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77f7885f7f-f7d66" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.073184 5018 scope.go:117] "RemoveContainer" containerID="bfd96cd797a9cef67b6577318e35f224798021a3e38496f28eb79e625754ac78" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.104768 5018 scope.go:117] "RemoveContainer" containerID="777701be4a1b35e6222e836b636046f71af77efe74c3e630f85c8067ea4696b4" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.122676 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.137580 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156152 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156526 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="init" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156541 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="init" Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156557 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156564 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156585 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156590 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156599 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156604 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156633 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="dnsmasq-dns" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156641 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="dnsmasq-dns" Oct 14 07:09:10 crc kubenswrapper[5018]: E1014 07:09:10.156649 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156656 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156814 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156827 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-httpd" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156841 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" containerName="dnsmasq-dns" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156852 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.156864 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" containerName="glance-log" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.157926 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.161719 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.162471 5018 scope.go:117] "RemoveContainer" containerID="b9b31d28ad5ef69f136cd455ee32490afde1ccd6e3aea1ed059cbe782593af82" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.162658 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.162800 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.167042 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pj22j" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.176075 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.200035 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77f7885f7f-f7d66"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.210938 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.217177 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.226118 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.246952 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247056 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247089 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247109 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247150 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247183 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jgfk\" (UniqueName: \"kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247211 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.247238 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.261710 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.263206 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.265840 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.266103 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.287087 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.313082 5018 scope.go:117] "RemoveContainer" containerID="d6662481997568702efbfc8fa1f8bfc82f15457b8a4d5500c6fb2edaf4f13b56" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.352409 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.355827 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.353467 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.355889 5018 scope.go:117] "RemoveContainer" containerID="f3f27d2120e9963fe5aec28385b0c7f1624e4a6ec9ed9738212db3391674a203" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356492 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356529 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356600 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356654 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356703 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jgfk\" (UniqueName: \"kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356740 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356776 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356885 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356950 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgjs8\" (UniqueName: \"kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356976 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.356995 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.357023 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.357049 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.357077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.357765 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.357975 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.359725 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.362583 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.363042 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.364576 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.386138 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jgfk\" (UniqueName: \"kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.395072 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458044 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458116 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458143 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458195 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgjs8\" (UniqueName: \"kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458220 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458248 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458265 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458284 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.458354 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.459297 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.460029 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.463610 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.464759 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.465175 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.465291 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.479612 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgjs8\" (UniqueName: \"kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.484898 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.609512 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.620778 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ecd3ea9-98af-4b53-a40a-8b5b1c48da41" path="/var/lib/kubelet/pods/0ecd3ea9-98af-4b53-a40a-8b5b1c48da41/volumes" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.621605 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44751d49-a17c-4e9e-b1df-e2710d5dc481" path="/var/lib/kubelet/pods/44751d49-a17c-4e9e-b1df-e2710d5dc481/volumes" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.622362 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5711950b-f891-4fc2-84fe-fd0c12d1b228" path="/var/lib/kubelet/pods/5711950b-f891-4fc2-84fe-fd0c12d1b228/volumes" Oct 14 07:09:10 crc kubenswrapper[5018]: I1014 07:09:10.631333 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.091914 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerStarted","Data":"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f"} Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.092342 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerStarted","Data":"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946"} Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.093353 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.093408 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.096363 5018 generic.go:334] "Generic (PLEG): container finished" podID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerID="e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d" exitCode=0 Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.096445 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5455955775-hs6f6" event={"ID":"7e54cbba-6800-47af-8d42-fcb95bcf621e","Type":"ContainerDied","Data":"e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d"} Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.112318 5018 generic.go:334] "Generic (PLEG): container finished" podID="2af870e0-db4f-49ba-afeb-238282c56c61" containerID="c32866d679f7b31297ab020b785b7625fa428a5926da8f4decfc634fc65e0844" exitCode=0 Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.112354 5018 generic.go:334] "Generic (PLEG): container finished" podID="2af870e0-db4f-49ba-afeb-238282c56c61" containerID="99b99075e3cc9efca3039a048d40e4309b1ee228733b7b47d0a3b9ce4f8f2730" exitCode=0 Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.112409 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerDied","Data":"c32866d679f7b31297ab020b785b7625fa428a5926da8f4decfc634fc65e0844"} Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.112444 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerDied","Data":"99b99075e3cc9efca3039a048d40e4309b1ee228733b7b47d0a3b9ce4f8f2730"} Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.135014 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bd6845fdb-hbp48" podStartSLOduration=3.134992071 podStartE2EDuration="3.134992071s" podCreationTimestamp="2025-10-14 07:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:11.126265894 +0000 UTC m=+1167.710312511" watchObservedRunningTime="2025-10-14 07:09:11.134992071 +0000 UTC m=+1167.719038698" Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.291582 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:11 crc kubenswrapper[5018]: I1014 07:09:11.350067 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:11 crc kubenswrapper[5018]: W1014 07:09:11.907718 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd2759b_9c06_4b72_8c05_a3b7617799f8.slice/crio-d9d718d74dc2da9a17195448db3e42f34a283ac80ee2b0153a12c6c656d6a1fb WatchSource:0}: Error finding container d9d718d74dc2da9a17195448db3e42f34a283ac80ee2b0153a12c6c656d6a1fb: Status 404 returned error can't find the container with id d9d718d74dc2da9a17195448db3e42f34a283ac80ee2b0153a12c6c656d6a1fb Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.136900 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerStarted","Data":"918408d26f8252f98ee0ce66986ce90d4d549d61f6ba6ec79e620053915595c1"} Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.141522 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerStarted","Data":"d9d718d74dc2da9a17195448db3e42f34a283ac80ee2b0153a12c6c656d6a1fb"} Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.372079 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.373628 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.379709 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.384181 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.390285 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512480 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w58jp\" (UniqueName: \"kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512558 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512612 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512657 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512694 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512721 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.512745 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.613976 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614079 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614127 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614182 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614224 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614260 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.614299 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w58jp\" (UniqueName: \"kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.615398 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.619910 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.620275 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.620239 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.620647 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.629066 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.629389 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w58jp\" (UniqueName: \"kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp\") pod \"barbican-api-dfd4968f6-zjg6m\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:12 crc kubenswrapper[5018]: I1014 07:09:12.698023 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.169180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerStarted","Data":"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.173604 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5455955775-hs6f6" event={"ID":"7e54cbba-6800-47af-8d42-fcb95bcf621e","Type":"ContainerStarted","Data":"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.178059 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.179438 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerStarted","Data":"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.179459 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerStarted","Data":"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.181966 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerStarted","Data":"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.181993 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerStarted","Data":"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.185086 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerStarted","Data":"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e"} Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.205987 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5455955775-hs6f6" podStartSLOduration=5.205970552 podStartE2EDuration="5.205970552s" podCreationTimestamp="2025-10-14 07:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:13.197037859 +0000 UTC m=+1169.781084486" watchObservedRunningTime="2025-10-14 07:09:13.205970552 +0000 UTC m=+1169.790017179" Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.233502 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" podStartSLOduration=3.129432774 podStartE2EDuration="5.233486532s" podCreationTimestamp="2025-10-14 07:09:08 +0000 UTC" firstStartedPulling="2025-10-14 07:09:09.861307448 +0000 UTC m=+1166.445354075" lastFinishedPulling="2025-10-14 07:09:11.965361206 +0000 UTC m=+1168.549407833" observedRunningTime="2025-10-14 07:09:13.211405196 +0000 UTC m=+1169.795451823" watchObservedRunningTime="2025-10-14 07:09:13.233486532 +0000 UTC m=+1169.817533159" Oct 14 07:09:13 crc kubenswrapper[5018]: W1014 07:09:13.257477 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75b77f89_1956_43e0_b6df_dd4d390e4cef.slice/crio-0ae60f1d8e72562e05a13c1170f1faa3e21e2b6e67f2b67664643d1b931f0f7a WatchSource:0}: Error finding container 0ae60f1d8e72562e05a13c1170f1faa3e21e2b6e67f2b67664643d1b931f0f7a: Status 404 returned error can't find the container with id 0ae60f1d8e72562e05a13c1170f1faa3e21e2b6e67f2b67664643d1b931f0f7a Oct 14 07:09:13 crc kubenswrapper[5018]: I1014 07:09:13.308544 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6d56c684b7-gtf99" podStartSLOduration=3.02337192 podStartE2EDuration="5.308525268s" podCreationTimestamp="2025-10-14 07:09:08 +0000 UTC" firstStartedPulling="2025-10-14 07:09:09.706592125 +0000 UTC m=+1166.290638772" lastFinishedPulling="2025-10-14 07:09:11.991745493 +0000 UTC m=+1168.575792120" observedRunningTime="2025-10-14 07:09:13.241785587 +0000 UTC m=+1169.825832214" watchObservedRunningTime="2025-10-14 07:09:13.308525268 +0000 UTC m=+1169.892571895" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.034690 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.195986 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerStarted","Data":"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.198581 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerStarted","Data":"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.198643 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerStarted","Data":"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.198659 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerStarted","Data":"0ae60f1d8e72562e05a13c1170f1faa3e21e2b6e67f2b67664643d1b931f0f7a"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.199198 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.199227 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.200317 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2b2jk" event={"ID":"25622217-99df-4f81-a099-b17af99a7dd6","Type":"ContainerStarted","Data":"19c6afa7b1ea3758086293b2bae47d5f7aba7eaf8fffdb5f13115ab7520de515"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.202724 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerStarted","Data":"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c"} Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.226373 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.226360249 podStartE2EDuration="4.226360249s" podCreationTimestamp="2025-10-14 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:14.223369725 +0000 UTC m=+1170.807416342" watchObservedRunningTime="2025-10-14 07:09:14.226360249 +0000 UTC m=+1170.810406876" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.252151 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-2b2jk" podStartSLOduration=5.39397781 podStartE2EDuration="37.252134799s" podCreationTimestamp="2025-10-14 07:08:37 +0000 UTC" firstStartedPulling="2025-10-14 07:08:41.359939567 +0000 UTC m=+1137.943986194" lastFinishedPulling="2025-10-14 07:09:13.218096556 +0000 UTC m=+1169.802143183" observedRunningTime="2025-10-14 07:09:14.245685597 +0000 UTC m=+1170.829732224" watchObservedRunningTime="2025-10-14 07:09:14.252134799 +0000 UTC m=+1170.836181426" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.269354 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.269337547 podStartE2EDuration="4.269337547s" podCreationTimestamp="2025-10-14 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:14.264757407 +0000 UTC m=+1170.848804034" watchObservedRunningTime="2025-10-14 07:09:14.269337547 +0000 UTC m=+1170.853384174" Oct 14 07:09:14 crc kubenswrapper[5018]: I1014 07:09:14.288980 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-dfd4968f6-zjg6m" podStartSLOduration=2.288963453 podStartE2EDuration="2.288963453s" podCreationTimestamp="2025-10-14 07:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:14.28427383 +0000 UTC m=+1170.868320457" watchObservedRunningTime="2025-10-14 07:09:14.288963453 +0000 UTC m=+1170.873010080" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.225927 5018 generic.go:334] "Generic (PLEG): container finished" podID="2af870e0-db4f-49ba-afeb-238282c56c61" containerID="d17b1e38a0f52e8c47fd151211bfd4af434143680579ee42c649929b908a86a0" exitCode=0 Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.226148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerDied","Data":"d17b1e38a0f52e8c47fd151211bfd4af434143680579ee42c649929b908a86a0"} Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.422818 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.466936 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467020 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8rsx\" (UniqueName: \"kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467084 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467146 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467270 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467313 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.467401 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data\") pod \"2af870e0-db4f-49ba-afeb-238282c56c61\" (UID: \"2af870e0-db4f-49ba-afeb-238282c56c61\") " Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.468030 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.468460 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.473230 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx" (OuterVolumeSpecName: "kube-api-access-s8rsx") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "kube-api-access-s8rsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.473513 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.473588 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts" (OuterVolumeSpecName: "scripts") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.569787 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.569833 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8rsx\" (UniqueName: \"kubernetes.io/projected/2af870e0-db4f-49ba-afeb-238282c56c61-kube-api-access-s8rsx\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.569848 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.569859 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.569871 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2af870e0-db4f-49ba-afeb-238282c56c61-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.589840 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.606381 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data" (OuterVolumeSpecName: "config-data") pod "2af870e0-db4f-49ba-afeb-238282c56c61" (UID: "2af870e0-db4f-49ba-afeb-238282c56c61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.671864 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:15 crc kubenswrapper[5018]: I1014 07:09:15.671899 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af870e0-db4f-49ba-afeb-238282c56c61-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.240429 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.240505 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2af870e0-db4f-49ba-afeb-238282c56c61","Type":"ContainerDied","Data":"80121411d168fc7e2f3e52c9ebe3811af90e9eb7ba2616ad2409acf0eb39acc7"} Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.240935 5018 scope.go:117] "RemoveContainer" containerID="c32866d679f7b31297ab020b785b7625fa428a5926da8f4decfc634fc65e0844" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.321647 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.333667 5018 scope.go:117] "RemoveContainer" containerID="d17b1e38a0f52e8c47fd151211bfd4af434143680579ee42c649929b908a86a0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.345291 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.350572 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:16 crc kubenswrapper[5018]: E1014 07:09:16.351032 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-notification-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351052 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-notification-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: E1014 07:09:16.351085 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-central-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351095 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-central-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: E1014 07:09:16.351108 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="proxy-httpd" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351115 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="proxy-httpd" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351323 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-central-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351347 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="proxy-httpd" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.351362 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" containerName="ceilometer-notification-agent" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.353383 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.356201 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.356298 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.364002 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.365932 5018 scope.go:117] "RemoveContainer" containerID="99b99075e3cc9efca3039a048d40e4309b1ee228733b7b47d0a3b9ce4f8f2730" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.383727 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.383906 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.384014 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2lbm\" (UniqueName: \"kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.384131 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.384234 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.384363 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.384488 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485395 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485452 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485484 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485507 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2lbm\" (UniqueName: \"kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485553 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485588 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.485660 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.487225 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.489213 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.492019 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.492086 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.492403 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.502336 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.505566 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2lbm\" (UniqueName: \"kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm\") pod \"ceilometer-0\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " pod="openstack/ceilometer-0" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.618826 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2af870e0-db4f-49ba-afeb-238282c56c61" path="/var/lib/kubelet/pods/2af870e0-db4f-49ba-afeb-238282c56c61/volumes" Oct 14 07:09:16 crc kubenswrapper[5018]: I1014 07:09:16.676690 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:17 crc kubenswrapper[5018]: I1014 07:09:17.222237 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:17 crc kubenswrapper[5018]: I1014 07:09:17.254245 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerStarted","Data":"33f40076df59cfcb0b6d0361033c0236783e33388134a9d4807873db273628a0"} Oct 14 07:09:18 crc kubenswrapper[5018]: I1014 07:09:18.266009 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerStarted","Data":"7fd79b97dcdf57e46266fcbc4e307abb85f811a686b29ba6b8480915b694e9bf"} Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.036231 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.087208 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.087491 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="dnsmasq-dns" containerID="cri-o://9df84315e41ea8b5a5e789dd8ace05732ab56f75aebfeec16c62b24045e65329" gracePeriod=10 Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.330822 5018 generic.go:334] "Generic (PLEG): container finished" podID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerID="9df84315e41ea8b5a5e789dd8ace05732ab56f75aebfeec16c62b24045e65329" exitCode=0 Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.330906 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" event={"ID":"8786087b-cf37-429a-9c1b-bd057f422c3d","Type":"ContainerDied","Data":"9df84315e41ea8b5a5e789dd8ace05732ab56f75aebfeec16c62b24045e65329"} Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.354313 5018 generic.go:334] "Generic (PLEG): container finished" podID="53363d98-901b-4334-bfcc-a74638255459" containerID="a69669374cda679bf77ce540f42dd59f2d251b3623db67fb8c294ce9d28fd5b8" exitCode=0 Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.354395 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nx7zt" event={"ID":"53363d98-901b-4334-bfcc-a74638255459","Type":"ContainerDied","Data":"a69669374cda679bf77ce540f42dd59f2d251b3623db67fb8c294ce9d28fd5b8"} Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.379818 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerStarted","Data":"24d86d1fc0be34055cd6378df65e13f2f30420de13c4afe3bea2392b1d5e778a"} Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.442054 5018 generic.go:334] "Generic (PLEG): container finished" podID="25622217-99df-4f81-a099-b17af99a7dd6" containerID="19c6afa7b1ea3758086293b2bae47d5f7aba7eaf8fffdb5f13115ab7520de515" exitCode=0 Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.442115 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2b2jk" event={"ID":"25622217-99df-4f81-a099-b17af99a7dd6","Type":"ContainerDied","Data":"19c6afa7b1ea3758086293b2bae47d5f7aba7eaf8fffdb5f13115ab7520de515"} Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.830888 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865533 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865657 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865758 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865796 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865816 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.865885 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58g7k\" (UniqueName: \"kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k\") pod \"8786087b-cf37-429a-9c1b-bd057f422c3d\" (UID: \"8786087b-cf37-429a-9c1b-bd057f422c3d\") " Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.886873 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k" (OuterVolumeSpecName: "kube-api-access-58g7k") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "kube-api-access-58g7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.960732 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.969344 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58g7k\" (UniqueName: \"kubernetes.io/projected/8786087b-cf37-429a-9c1b-bd057f422c3d-kube-api-access-58g7k\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.969379 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.976288 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.996203 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:19 crc kubenswrapper[5018]: I1014 07:09:19.997241 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config" (OuterVolumeSpecName: "config") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.023977 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8786087b-cf37-429a-9c1b-bd057f422c3d" (UID: "8786087b-cf37-429a-9c1b-bd057f422c3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.070939 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.070971 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.070981 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.070995 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8786087b-cf37-429a-9c1b-bd057f422c3d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.465972 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" event={"ID":"8786087b-cf37-429a-9c1b-bd057f422c3d","Type":"ContainerDied","Data":"d68f2454841be95f2ee71cceb8da693e91928c0be731adcd3ffe6f6b6779a6cd"} Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.466017 5018 scope.go:117] "RemoveContainer" containerID="9df84315e41ea8b5a5e789dd8ace05732ab56f75aebfeec16c62b24045e65329" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.466122 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.490925 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerStarted","Data":"084a9c2a21923f20e85e8d9aca17be12df497b5b8a50604239e63426973d19eb"} Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.519177 5018 scope.go:117] "RemoveContainer" containerID="3a3ea19ac068e25ad6df8b649097f8de9ef2b573ca06c17189eba6c123e13c2b" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.524887 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.532999 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c75bdf7bf-x9gtp"] Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.640749 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" path="/var/lib/kubelet/pods/8786087b-cf37-429a-9c1b-bd057f422c3d/volumes" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.641461 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.641485 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.641495 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.641503 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.661181 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.663159 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.696212 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.708589 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.869081 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.869130 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:09:20 crc kubenswrapper[5018]: I1014 07:09:20.908844 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.019706 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.090505 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.090722 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.091360 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.091402 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.091693 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.091843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szr4x\" (UniqueName: \"kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.091886 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data\") pod \"25622217-99df-4f81-a099-b17af99a7dd6\" (UID: \"25622217-99df-4f81-a099-b17af99a7dd6\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.092234 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/25622217-99df-4f81-a099-b17af99a7dd6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.099849 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.109059 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts" (OuterVolumeSpecName: "scripts") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.109234 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x" (OuterVolumeSpecName: "kube-api-access-szr4x") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "kube-api-access-szr4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.135860 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.159255 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data" (OuterVolumeSpecName: "config-data") pod "25622217-99df-4f81-a099-b17af99a7dd6" (UID: "25622217-99df-4f81-a099-b17af99a7dd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.192796 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config\") pod \"53363d98-901b-4334-bfcc-a74638255459\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.192938 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhvbc\" (UniqueName: \"kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc\") pod \"53363d98-901b-4334-bfcc-a74638255459\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193058 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle\") pod \"53363d98-901b-4334-bfcc-a74638255459\" (UID: \"53363d98-901b-4334-bfcc-a74638255459\") " Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193387 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szr4x\" (UniqueName: \"kubernetes.io/projected/25622217-99df-4f81-a099-b17af99a7dd6-kube-api-access-szr4x\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193403 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193414 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193422 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.193430 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25622217-99df-4f81-a099-b17af99a7dd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.201477 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc" (OuterVolumeSpecName: "kube-api-access-hhvbc") pod "53363d98-901b-4334-bfcc-a74638255459" (UID: "53363d98-901b-4334-bfcc-a74638255459"). InnerVolumeSpecName "kube-api-access-hhvbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.209819 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.215402 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config" (OuterVolumeSpecName: "config") pod "53363d98-901b-4334-bfcc-a74638255459" (UID: "53363d98-901b-4334-bfcc-a74638255459"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.230521 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53363d98-901b-4334-bfcc-a74638255459" (UID: "53363d98-901b-4334-bfcc-a74638255459"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.294513 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.294548 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhvbc\" (UniqueName: \"kubernetes.io/projected/53363d98-901b-4334-bfcc-a74638255459-kube-api-access-hhvbc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.294562 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53363d98-901b-4334-bfcc-a74638255459-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.500013 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerStarted","Data":"6bcb6d325818100d136b12555d42017c0070b89b74b89f2e51e92446c7d8cb3e"} Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.501303 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.502546 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2b2jk" event={"ID":"25622217-99df-4f81-a099-b17af99a7dd6","Type":"ContainerDied","Data":"213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508"} Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.502569 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="213b3901a672c9a69a7c8c62883084f0f55c74dca909a85623e2d7c5e42d0508" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.502612 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2b2jk" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.519973 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nx7zt" event={"ID":"53363d98-901b-4334-bfcc-a74638255459","Type":"ContainerDied","Data":"6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4"} Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520007 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ea332d8e18a8280762e97f531419f024ef1b8772904464e6483b0a2d26a3cc4" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520048 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nx7zt" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520585 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520696 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520713 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.520722 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.525038 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.561061 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.652267697 podStartE2EDuration="5.561046593s" podCreationTimestamp="2025-10-14 07:09:16 +0000 UTC" firstStartedPulling="2025-10-14 07:09:17.219544338 +0000 UTC m=+1173.803590965" lastFinishedPulling="2025-10-14 07:09:21.128323234 +0000 UTC m=+1177.712369861" observedRunningTime="2025-10-14 07:09:21.55776825 +0000 UTC m=+1178.141814877" watchObservedRunningTime="2025-10-14 07:09:21.561046593 +0000 UTC m=+1178.145093220" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612113 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:21 crc kubenswrapper[5018]: E1014 07:09:21.612461 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="dnsmasq-dns" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612476 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="dnsmasq-dns" Oct 14 07:09:21 crc kubenswrapper[5018]: E1014 07:09:21.612495 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="init" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612502 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="init" Oct 14 07:09:21 crc kubenswrapper[5018]: E1014 07:09:21.612520 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25622217-99df-4f81-a099-b17af99a7dd6" containerName="cinder-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612526 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="25622217-99df-4f81-a099-b17af99a7dd6" containerName="cinder-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: E1014 07:09:21.612541 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53363d98-901b-4334-bfcc-a74638255459" containerName="neutron-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612547 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="53363d98-901b-4334-bfcc-a74638255459" containerName="neutron-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612721 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="53363d98-901b-4334-bfcc-a74638255459" containerName="neutron-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612734 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="dnsmasq-dns" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.612745 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="25622217-99df-4f81-a099-b17af99a7dd6" containerName="cinder-db-sync" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.613574 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.644427 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700196 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700237 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqspm\" (UniqueName: \"kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700321 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700420 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700484 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.700520 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.736448 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.737920 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.745286 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.750541 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.750582 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.750743 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-k7ccr" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.763645 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801534 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801592 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqspm\" (UniqueName: \"kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801629 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801653 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801695 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801715 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6q8r\" (UniqueName: \"kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801737 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801777 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801798 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801822 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.801859 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.802709 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.802738 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.803249 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.803423 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.814250 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.828765 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqspm\" (UniqueName: \"kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm\") pod \"dnsmasq-dns-6cfb6b659f-sjd6p\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.851841 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.872013 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.879021 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8r9xt" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.879223 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.879398 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.879500 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.895097 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903388 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903433 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903485 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903512 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903542 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903579 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szjgs\" (UniqueName: \"kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903597 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903627 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903646 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6q8r\" (UniqueName: \"kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903698 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.903727 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.921756 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.925193 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.925784 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.926576 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.954164 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:21 crc kubenswrapper[5018]: I1014 07:09:21.960224 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6q8r\" (UniqueName: \"kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r\") pod \"neutron-7bc7b47c6d-sbkf8\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013507 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013565 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013583 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013635 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013678 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szjgs\" (UniqueName: \"kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.013693 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.015306 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.025131 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.031182 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.032817 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.034099 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.035160 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.084157 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szjgs\" (UniqueName: \"kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs\") pod \"cinder-scheduler-0\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.091430 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.092776 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.093771 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.146642 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.221852 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.221930 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.221955 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.221987 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.222018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.222038 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf8lj\" (UniqueName: \"kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.225492 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.233705 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.244485 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.246276 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331604 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2thw\" (UniqueName: \"kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331662 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331689 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331761 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331777 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331796 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331817 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331877 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331891 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331907 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.331928 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf8lj\" (UniqueName: \"kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.332959 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.333257 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.333534 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.333809 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.334080 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.362831 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf8lj\" (UniqueName: \"kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj\") pod \"dnsmasq-dns-cb9f44c77-w69fd\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.364025 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.432825 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2thw\" (UniqueName: \"kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433034 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433094 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433110 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433132 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433178 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.433194 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.436364 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.437317 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.437556 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.439425 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.442103 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.447269 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.453123 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.472682 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2thw\" (UniqueName: \"kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw\") pod \"cinder-api-0\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.585226 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:22 crc kubenswrapper[5018]: I1014 07:09:22.667237 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:22 crc kubenswrapper[5018]: W1014 07:09:22.698477 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17c323aa_493e_4bdd_8dfb_165c867e68a5.slice/crio-1cbc903c3a279741c2970681ec670e98dd6cb7a6bb2f42a60cf5393d9f82bd77 WatchSource:0}: Error finding container 1cbc903c3a279741c2970681ec670e98dd6cb7a6bb2f42a60cf5393d9f82bd77: Status 404 returned error can't find the container with id 1cbc903c3a279741c2970681ec670e98dd6cb7a6bb2f42a60cf5393d9f82bd77 Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.127756 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.135738 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:09:23 crc kubenswrapper[5018]: W1014 07:09:23.142512 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfee0da5_22e6_49c5_988c_de2781a8dd53.slice/crio-80098bcadd126150546df569f827c8db720939e4ba0febf6a8b445df434a5ce8 WatchSource:0}: Error finding container 80098bcadd126150546df569f827c8db720939e4ba0febf6a8b445df434a5ce8: Status 404 returned error can't find the container with id 80098bcadd126150546df569f827c8db720939e4ba0febf6a8b445df434a5ce8 Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.298654 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.317411 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.666997 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerStarted","Data":"988cefb3a1377486426dc94dc8edf03b6515c192b1f42bea18f62a241d986d19"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.695248 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" event={"ID":"67d09018-563a-4a79-aa5b-84ca6f7c441e","Type":"ContainerStarted","Data":"1d742d1b1a011113357dbe71e47d89043ec4cd1369d00dbf2923cac1735ca4f8"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.760823 5018 generic.go:334] "Generic (PLEG): container finished" podID="17c323aa-493e-4bdd-8dfb-165c867e68a5" containerID="4f5527d05a403096cddb594ecb180d9e5cc95ebbd137fb79405cfeaeaa2359b7" exitCode=0 Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.761009 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" event={"ID":"17c323aa-493e-4bdd-8dfb-165c867e68a5","Type":"ContainerDied","Data":"4f5527d05a403096cddb594ecb180d9e5cc95ebbd137fb79405cfeaeaa2359b7"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.761032 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" event={"ID":"17c323aa-493e-4bdd-8dfb-165c867e68a5","Type":"ContainerStarted","Data":"1cbc903c3a279741c2970681ec670e98dd6cb7a6bb2f42a60cf5393d9f82bd77"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.850839 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerStarted","Data":"80098bcadd126150546df569f827c8db720939e4ba0febf6a8b445df434a5ce8"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.873681 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerStarted","Data":"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.873729 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerStarted","Data":"b3a2e763c70f090a60ded9af125db1d8a8dea1208209fd50ed39ec58f8f5cbd9"} Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.873816 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:09:23 crc kubenswrapper[5018]: I1014 07:09:23.873825 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.301654 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413211 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413308 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413375 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413424 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqspm\" (UniqueName: \"kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413477 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.413494 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc\") pod \"17c323aa-493e-4bdd-8dfb-165c867e68a5\" (UID: \"17c323aa-493e-4bdd-8dfb-165c867e68a5\") " Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.456737 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.457118 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm" (OuterVolumeSpecName: "kube-api-access-nqspm") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "kube-api-access-nqspm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.497577 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.497748 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c75bdf7bf-x9gtp" podUID="8786087b-cf37-429a-9c1b-bd057f422c3d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: i/o timeout" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.500096 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config" (OuterVolumeSpecName: "config") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.515975 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.516005 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.516015 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.516023 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqspm\" (UniqueName: \"kubernetes.io/projected/17c323aa-493e-4bdd-8dfb-165c867e68a5-kube-api-access-nqspm\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.537758 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.623202 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.658193 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17c323aa-493e-4bdd-8dfb-165c867e68a5" (UID: "17c323aa-493e-4bdd-8dfb-165c867e68a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.744235 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17c323aa-493e-4bdd-8dfb-165c867e68a5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.907841 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" event={"ID":"17c323aa-493e-4bdd-8dfb-165c867e68a5","Type":"ContainerDied","Data":"1cbc903c3a279741c2970681ec670e98dd6cb7a6bb2f42a60cf5393d9f82bd77"} Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.907893 5018 scope.go:117] "RemoveContainer" containerID="4f5527d05a403096cddb594ecb180d9e5cc95ebbd137fb79405cfeaeaa2359b7" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.908025 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfb6b659f-sjd6p" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.932725 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerStarted","Data":"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40"} Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.933413 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.949816 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerStarted","Data":"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82"} Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.984168 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.984666 5018 generic.go:334] "Generic (PLEG): container finished" podID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerID="5d3f7f51b4162c7d00450085a977315fd391d7960354fba47bf2850f1ad13c4a" exitCode=0 Oct 14 07:09:24 crc kubenswrapper[5018]: I1014 07:09:24.984692 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" event={"ID":"67d09018-563a-4a79-aa5b-84ca6f7c441e","Type":"ContainerDied","Data":"5d3f7f51b4162c7d00450085a977315fd391d7960354fba47bf2850f1ad13c4a"} Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.020749 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cfb6b659f-sjd6p"] Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.027322 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bc7b47c6d-sbkf8" podStartSLOduration=4.027300153 podStartE2EDuration="4.027300153s" podCreationTimestamp="2025-10-14 07:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:24.991989523 +0000 UTC m=+1181.576036150" watchObservedRunningTime="2025-10-14 07:09:25.027300153 +0000 UTC m=+1181.611346780" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.052987 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.183821 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.184169 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.259403 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.882324 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.972786 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.973165 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bd6845fdb-hbp48" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api-log" containerID="cri-o://fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946" gracePeriod=30 Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.973470 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bd6845fdb-hbp48" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api" containerID="cri-o://24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f" gracePeriod=30 Oct 14 07:09:25 crc kubenswrapper[5018]: I1014 07:09:25.985908 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.033944 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerStarted","Data":"dab56388c6e4d47b8260023cc1f4193441ec8c3d7619d5bd3c822262d9315d87"} Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.056788 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" event={"ID":"67d09018-563a-4a79-aa5b-84ca6f7c441e","Type":"ContainerStarted","Data":"2d0c149e193aa09e131eec98a80bc6bb0be317bd1605f38db67dc9c18dd80398"} Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.057633 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.281221 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.281313 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.309172 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" podStartSLOduration=4.309155499 podStartE2EDuration="4.309155499s" podCreationTimestamp="2025-10-14 07:09:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:26.089700861 +0000 UTC m=+1182.673747488" watchObservedRunningTime="2025-10-14 07:09:26.309155499 +0000 UTC m=+1182.893202116" Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.614743 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17c323aa-493e-4bdd-8dfb-165c867e68a5" path="/var/lib/kubelet/pods/17c323aa-493e-4bdd-8dfb-165c867e68a5/volumes" Oct 14 07:09:26 crc kubenswrapper[5018]: I1014 07:09:26.760599 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.065277 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerStarted","Data":"8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8"} Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.067043 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerStarted","Data":"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f"} Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.067213 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api-log" containerID="cri-o://a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" gracePeriod=30 Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.067225 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.067231 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api" containerID="cri-o://e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" gracePeriod=30 Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.069605 5018 generic.go:334] "Generic (PLEG): container finished" podID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerID="fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946" exitCode=143 Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.070431 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerDied","Data":"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946"} Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.089403 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.903401034 podStartE2EDuration="6.089377483s" podCreationTimestamp="2025-10-14 07:09:21 +0000 UTC" firstStartedPulling="2025-10-14 07:09:23.179430213 +0000 UTC m=+1179.763476840" lastFinishedPulling="2025-10-14 07:09:24.365406672 +0000 UTC m=+1180.949453289" observedRunningTime="2025-10-14 07:09:27.083591629 +0000 UTC m=+1183.667638256" watchObservedRunningTime="2025-10-14 07:09:27.089377483 +0000 UTC m=+1183.673424120" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.113899 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.113881687 podStartE2EDuration="5.113881687s" podCreationTimestamp="2025-10-14 07:09:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:27.110127701 +0000 UTC m=+1183.694174348" watchObservedRunningTime="2025-10-14 07:09:27.113881687 +0000 UTC m=+1183.697928314" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.364887 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.733164 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809094 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809142 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809182 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809263 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2thw\" (UniqueName: \"kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809282 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809316 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.809373 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom\") pod \"74617685-4df1-48ad-bf44-fc473323d0b8\" (UID: \"74617685-4df1-48ad-bf44-fc473323d0b8\") " Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.811742 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.814938 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs" (OuterVolumeSpecName: "logs") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.827700 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.833024 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts" (OuterVolumeSpecName: "scripts") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.842780 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw" (OuterVolumeSpecName: "kube-api-access-s2thw") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "kube-api-access-s2thw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.859542 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911781 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2thw\" (UniqueName: \"kubernetes.io/projected/74617685-4df1-48ad-bf44-fc473323d0b8-kube-api-access-s2thw\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911816 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911825 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74617685-4df1-48ad-bf44-fc473323d0b8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911834 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911848 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74617685-4df1-48ad-bf44-fc473323d0b8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.911856 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:27 crc kubenswrapper[5018]: I1014 07:09:27.920983 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data" (OuterVolumeSpecName: "config-data") pod "74617685-4df1-48ad-bf44-fc473323d0b8" (UID: "74617685-4df1-48ad-bf44-fc473323d0b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.013433 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74617685-4df1-48ad-bf44-fc473323d0b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080009 5018 generic.go:334] "Generic (PLEG): container finished" podID="74617685-4df1-48ad-bf44-fc473323d0b8" containerID="e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" exitCode=0 Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080044 5018 generic.go:334] "Generic (PLEG): container finished" podID="74617685-4df1-48ad-bf44-fc473323d0b8" containerID="a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" exitCode=143 Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080117 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerDied","Data":"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f"} Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080160 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerDied","Data":"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82"} Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080174 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"74617685-4df1-48ad-bf44-fc473323d0b8","Type":"ContainerDied","Data":"988cefb3a1377486426dc94dc8edf03b6515c192b1f42bea18f62a241d986d19"} Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.080191 5018 scope.go:117] "RemoveContainer" containerID="e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.081715 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.107002 5018 scope.go:117] "RemoveContainer" containerID="a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.118422 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.126138 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.132278 5018 scope.go:117] "RemoveContainer" containerID="e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" Oct 14 07:09:28 crc kubenswrapper[5018]: E1014 07:09:28.132749 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f\": container with ID starting with e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f not found: ID does not exist" containerID="e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.132778 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f"} err="failed to get container status \"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f\": rpc error: code = NotFound desc = could not find container \"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f\": container with ID starting with e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f not found: ID does not exist" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.132799 5018 scope.go:117] "RemoveContainer" containerID="a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" Oct 14 07:09:28 crc kubenswrapper[5018]: E1014 07:09:28.135851 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82\": container with ID starting with a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82 not found: ID does not exist" containerID="a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.135902 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82"} err="failed to get container status \"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82\": rpc error: code = NotFound desc = could not find container \"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82\": container with ID starting with a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82 not found: ID does not exist" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.135928 5018 scope.go:117] "RemoveContainer" containerID="e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.136371 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f"} err="failed to get container status \"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f\": rpc error: code = NotFound desc = could not find container \"e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f\": container with ID starting with e1146ed2adf2a356847e09dc45d574f6a4e8125e2be7d32d91908f851d00bd9f not found: ID does not exist" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.136418 5018 scope.go:117] "RemoveContainer" containerID="a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.139507 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82"} err="failed to get container status \"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82\": rpc error: code = NotFound desc = could not find container \"a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82\": container with ID starting with a8f1763c56e66539788de5163c9eef350b13cc2dba6773e052721e65c5ec4a82 not found: ID does not exist" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.146628 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:28 crc kubenswrapper[5018]: E1014 07:09:28.146924 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c323aa-493e-4bdd-8dfb-165c867e68a5" containerName="init" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.146941 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c323aa-493e-4bdd-8dfb-165c867e68a5" containerName="init" Oct 14 07:09:28 crc kubenswrapper[5018]: E1014 07:09:28.146966 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.146973 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api" Oct 14 07:09:28 crc kubenswrapper[5018]: E1014 07:09:28.146988 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api-log" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.146995 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api-log" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.147176 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.147192 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" containerName="cinder-api-log" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.147207 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c323aa-493e-4bdd-8dfb-165c867e68a5" containerName="init" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.148405 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.150916 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.151133 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.151194 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.167440 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.217357 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqrsh\" (UniqueName: \"kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.217559 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.217779 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.217895 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.218040 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.218077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.218112 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.218135 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.218188 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.277713 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.279247 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.281660 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.286739 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.289220 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319337 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319597 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319660 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319679 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319701 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319726 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319748 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319768 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319796 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319830 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319852 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrsh\" (UniqueName: \"kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319871 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr9pc\" (UniqueName: \"kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319905 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319923 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.319954 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.320428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.320916 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.325111 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.328142 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.328210 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.328652 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.332218 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.340764 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqrsh\" (UniqueName: \"kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.346506 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom\") pod \"cinder-api-0\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421467 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421516 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr9pc\" (UniqueName: \"kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421536 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421553 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421602 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421667 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.421700 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.424904 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.425356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.426412 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.427971 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.428274 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.428726 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.447657 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr9pc\" (UniqueName: \"kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc\") pod \"neutron-75695fb6c7-7ztlf\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.468251 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.609746 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.622140 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74617685-4df1-48ad-bf44-fc473323d0b8" path="/var/lib/kubelet/pods/74617685-4df1-48ad-bf44-fc473323d0b8/volumes" Oct 14 07:09:28 crc kubenswrapper[5018]: I1014 07:09:28.918416 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.102295 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerStarted","Data":"94af7b16c95a3ae7f24f31c266c8993c1adb8a361512fbf5e0be6ee10506b46d"} Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.152486 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.414418 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bd6845fdb-hbp48" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:55136->10.217.0.154:9311: read: connection reset by peer" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.414454 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bd6845fdb-hbp48" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:55140->10.217.0.154:9311: read: connection reset by peer" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.913917 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.950008 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data\") pod \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.950084 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d77vf\" (UniqueName: \"kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf\") pod \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.950219 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom\") pod \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.950251 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs\") pod \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.950470 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle\") pod \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\" (UID: \"c8954b74-2dcc-45bc-ad3f-205bdd3de472\") " Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.955841 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs" (OuterVolumeSpecName: "logs") pod "c8954b74-2dcc-45bc-ad3f-205bdd3de472" (UID: "c8954b74-2dcc-45bc-ad3f-205bdd3de472"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.970873 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c8954b74-2dcc-45bc-ad3f-205bdd3de472" (UID: "c8954b74-2dcc-45bc-ad3f-205bdd3de472"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.971022 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf" (OuterVolumeSpecName: "kube-api-access-d77vf") pod "c8954b74-2dcc-45bc-ad3f-205bdd3de472" (UID: "c8954b74-2dcc-45bc-ad3f-205bdd3de472"). InnerVolumeSpecName "kube-api-access-d77vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:29 crc kubenswrapper[5018]: I1014 07:09:29.991772 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8954b74-2dcc-45bc-ad3f-205bdd3de472" (UID: "c8954b74-2dcc-45bc-ad3f-205bdd3de472"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.014770 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data" (OuterVolumeSpecName: "config-data") pod "c8954b74-2dcc-45bc-ad3f-205bdd3de472" (UID: "c8954b74-2dcc-45bc-ad3f-205bdd3de472"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.053711 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.053741 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.053750 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d77vf\" (UniqueName: \"kubernetes.io/projected/c8954b74-2dcc-45bc-ad3f-205bdd3de472-kube-api-access-d77vf\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.053760 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8954b74-2dcc-45bc-ad3f-205bdd3de472-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.053770 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8954b74-2dcc-45bc-ad3f-205bdd3de472-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.115780 5018 generic.go:334] "Generic (PLEG): container finished" podID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerID="24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f" exitCode=0 Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.115862 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bd6845fdb-hbp48" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.115867 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerDied","Data":"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.115991 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bd6845fdb-hbp48" event={"ID":"c8954b74-2dcc-45bc-ad3f-205bdd3de472","Type":"ContainerDied","Data":"3d67473c73ce5ff0b49ff7efeeb23d11ea90445d253f7e91ae84e9887ff6527e"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.116013 5018 scope.go:117] "RemoveContainer" containerID="24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.119552 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerStarted","Data":"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.130905 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerStarted","Data":"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.130949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerStarted","Data":"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.130960 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerStarted","Data":"5a1f482ad48c64cd1008be1b1ffec437fe4b450cfba9ff6cf964744b364053e6"} Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.131809 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.141279 5018 scope.go:117] "RemoveContainer" containerID="fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.164023 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-75695fb6c7-7ztlf" podStartSLOduration=2.164007028 podStartE2EDuration="2.164007028s" podCreationTimestamp="2025-10-14 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:30.15914231 +0000 UTC m=+1186.743188947" watchObservedRunningTime="2025-10-14 07:09:30.164007028 +0000 UTC m=+1186.748053655" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.176814 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.181162 5018 scope.go:117] "RemoveContainer" containerID="24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f" Oct 14 07:09:30 crc kubenswrapper[5018]: E1014 07:09:30.181760 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f\": container with ID starting with 24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f not found: ID does not exist" containerID="24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.181806 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f"} err="failed to get container status \"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f\": rpc error: code = NotFound desc = could not find container \"24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f\": container with ID starting with 24f5970738ddc37d21d31558fcf9ff13f1e3556b754c8fa19b010410743ecb3f not found: ID does not exist" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.181852 5018 scope.go:117] "RemoveContainer" containerID="fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946" Oct 14 07:09:30 crc kubenswrapper[5018]: E1014 07:09:30.182246 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946\": container with ID starting with fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946 not found: ID does not exist" containerID="fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.182271 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946"} err="failed to get container status \"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946\": rpc error: code = NotFound desc = could not find container \"fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946\": container with ID starting with fc493657adcae624200d62d6353405b54e8762f58a9766b07d773fba1aaf1946 not found: ID does not exist" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.184187 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7bd6845fdb-hbp48"] Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.628983 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" path="/var/lib/kubelet/pods/c8954b74-2dcc-45bc-ad3f-205bdd3de472/volumes" Oct 14 07:09:30 crc kubenswrapper[5018]: I1014 07:09:30.967466 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:09:31 crc kubenswrapper[5018]: I1014 07:09:31.152682 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerStarted","Data":"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6"} Oct 14 07:09:31 crc kubenswrapper[5018]: I1014 07:09:31.153031 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 07:09:31 crc kubenswrapper[5018]: I1014 07:09:31.181098 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.181076812 podStartE2EDuration="3.181076812s" podCreationTimestamp="2025-10-14 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:31.172199001 +0000 UTC m=+1187.756245638" watchObservedRunningTime="2025-10-14 07:09:31.181076812 +0000 UTC m=+1187.765123439" Oct 14 07:09:32 crc kubenswrapper[5018]: I1014 07:09:32.441437 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:09:32 crc kubenswrapper[5018]: I1014 07:09:32.522819 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:32 crc kubenswrapper[5018]: I1014 07:09:32.523103 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5455955775-hs6f6" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="dnsmasq-dns" containerID="cri-o://04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d" gracePeriod=10 Oct 14 07:09:32 crc kubenswrapper[5018]: I1014 07:09:32.669606 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 07:09:32 crc kubenswrapper[5018]: I1014 07:09:32.739743 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.042961 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107346 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107399 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107472 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107509 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.107527 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zchgt\" (UniqueName: \"kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt\") pod \"7e54cbba-6800-47af-8d42-fcb95bcf621e\" (UID: \"7e54cbba-6800-47af-8d42-fcb95bcf621e\") " Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.113199 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt" (OuterVolumeSpecName: "kube-api-access-zchgt") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "kube-api-access-zchgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.155158 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.157642 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.163729 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config" (OuterVolumeSpecName: "config") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178097 5018 generic.go:334] "Generic (PLEG): container finished" podID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerID="04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d" exitCode=0 Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178236 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5455955775-hs6f6" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178290 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5455955775-hs6f6" event={"ID":"7e54cbba-6800-47af-8d42-fcb95bcf621e","Type":"ContainerDied","Data":"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d"} Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178354 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5455955775-hs6f6" event={"ID":"7e54cbba-6800-47af-8d42-fcb95bcf621e","Type":"ContainerDied","Data":"8e105987145d656de6eed8c37a3020c3937a4492e628cec57be189dd9636dc2c"} Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178373 5018 scope.go:117] "RemoveContainer" containerID="04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178398 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="cinder-scheduler" containerID="cri-o://dab56388c6e4d47b8260023cc1f4193441ec8c3d7619d5bd3c822262d9315d87" gracePeriod=30 Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.178751 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="probe" containerID="cri-o://8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8" gracePeriod=30 Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.188868 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e54cbba-6800-47af-8d42-fcb95bcf621e" (UID: "7e54cbba-6800-47af-8d42-fcb95bcf621e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209879 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209906 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209915 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209927 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209936 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e54cbba-6800-47af-8d42-fcb95bcf621e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.209944 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zchgt\" (UniqueName: \"kubernetes.io/projected/7e54cbba-6800-47af-8d42-fcb95bcf621e-kube-api-access-zchgt\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.237493 5018 scope.go:117] "RemoveContainer" containerID="e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.255920 5018 scope.go:117] "RemoveContainer" containerID="04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d" Oct 14 07:09:33 crc kubenswrapper[5018]: E1014 07:09:33.256414 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d\": container with ID starting with 04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d not found: ID does not exist" containerID="04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.256459 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d"} err="failed to get container status \"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d\": rpc error: code = NotFound desc = could not find container \"04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d\": container with ID starting with 04b963d836203e60bf6a6d8defa61d41ec24704e5fda1ceb8b6f4d6ff687c60d not found: ID does not exist" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.256484 5018 scope.go:117] "RemoveContainer" containerID="e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d" Oct 14 07:09:33 crc kubenswrapper[5018]: E1014 07:09:33.256874 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d\": container with ID starting with e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d not found: ID does not exist" containerID="e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.256907 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d"} err="failed to get container status \"e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d\": rpc error: code = NotFound desc = could not find container \"e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d\": container with ID starting with e8f4685d2682ce2bcad5bce58d2a5eec91f620c9b618ae0a6a11bfda73c3ce0d not found: ID does not exist" Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.511001 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:33 crc kubenswrapper[5018]: I1014 07:09:33.519472 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5455955775-hs6f6"] Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.032300 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfee0da5_22e6_49c5_988c_de2781a8dd53.slice/crio-8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.190723 5018 generic.go:334] "Generic (PLEG): container finished" podID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerID="8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8" exitCode=0 Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.190791 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerDied","Data":"8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8"} Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.638536 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" path="/var/lib/kubelet/pods/7e54cbba-6800-47af-8d42-fcb95bcf621e/volumes" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.650664 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.651101 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="dnsmasq-dns" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651122 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="dnsmasq-dns" Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.651136 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api-log" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651142 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api-log" Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.651154 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="init" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651160 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="init" Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.651175 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651180 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651357 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api-log" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651376 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e54cbba-6800-47af-8d42-fcb95bcf621e" containerName="dnsmasq-dns" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651387 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8954b74-2dcc-45bc-ad3f-205bdd3de472" containerName="barbican-api" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.651988 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.654459 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.654537 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.654679 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zc2x8" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.662580 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.739633 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.740045 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsvmj\" (UniqueName: \"kubernetes.io/projected/1954e1a7-b433-4ad3-a76a-2147db4b56c1-kube-api-access-nsvmj\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.740122 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.740152 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.813014 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.813776 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-nsvmj openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="1954e1a7-b433-4ad3-a76a-2147db4b56c1" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.821297 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.841716 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.841786 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.841862 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.841951 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsvmj\" (UniqueName: \"kubernetes.io/projected/1954e1a7-b433-4ad3-a76a-2147db4b56c1-kube-api-access-nsvmj\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.842897 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.843643 5018 projected.go:194] Error preparing data for projected volume kube-api-access-nsvmj for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 14 07:09:34 crc kubenswrapper[5018]: E1014 07:09:34.843708 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1954e1a7-b433-4ad3-a76a-2147db4b56c1-kube-api-access-nsvmj podName:1954e1a7-b433-4ad3-a76a-2147db4b56c1 nodeName:}" failed. No retries permitted until 2025-10-14 07:09:35.343689225 +0000 UTC m=+1191.927735852 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nsvmj" (UniqueName: "kubernetes.io/projected/1954e1a7-b433-4ad3-a76a-2147db4b56c1-kube-api-access-nsvmj") pod "openstackclient" (UID: "1954e1a7-b433-4ad3-a76a-2147db4b56c1") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.848063 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.848125 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.913389 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.914763 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.943333 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.943397 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhw2d\" (UniqueName: \"kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.943451 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.943580 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:34 crc kubenswrapper[5018]: I1014 07:09:34.946066 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.045017 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.045116 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.045142 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhw2d\" (UniqueName: \"kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.045177 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.045952 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.049868 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.050248 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.064778 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhw2d\" (UniqueName: \"kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d\") pod \"openstackclient\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.202569 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.208197 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1954e1a7-b433-4ad3-a76a-2147db4b56c1" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.227993 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.247581 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config\") pod \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.247767 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle\") pod \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.247850 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret\") pod \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\" (UID: \"1954e1a7-b433-4ad3-a76a-2147db4b56c1\") " Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.248026 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1954e1a7-b433-4ad3-a76a-2147db4b56c1" (UID: "1954e1a7-b433-4ad3-a76a-2147db4b56c1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.248454 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.248512 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsvmj\" (UniqueName: \"kubernetes.io/projected/1954e1a7-b433-4ad3-a76a-2147db4b56c1-kube-api-access-nsvmj\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.264588 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.290520 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1954e1a7-b433-4ad3-a76a-2147db4b56c1" (UID: "1954e1a7-b433-4ad3-a76a-2147db4b56c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.290927 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1954e1a7-b433-4ad3-a76a-2147db4b56c1" (UID: "1954e1a7-b433-4ad3-a76a-2147db4b56c1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.350493 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.350902 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1954e1a7-b433-4ad3-a76a-2147db4b56c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:35 crc kubenswrapper[5018]: I1014 07:09:35.766833 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:09:35 crc kubenswrapper[5018]: W1014 07:09:35.770177 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1719e0be_9beb_4f35_b449_2a90e51ec80c.slice/crio-b720e0212d91ca849011feb9d2e8c33d39b17061b220ebc69b656cacb4156bbe WatchSource:0}: Error finding container b720e0212d91ca849011feb9d2e8c33d39b17061b220ebc69b656cacb4156bbe: Status 404 returned error can't find the container with id b720e0212d91ca849011feb9d2e8c33d39b17061b220ebc69b656cacb4156bbe Oct 14 07:09:36 crc kubenswrapper[5018]: I1014 07:09:36.210540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1719e0be-9beb-4f35-b449-2a90e51ec80c","Type":"ContainerStarted","Data":"b720e0212d91ca849011feb9d2e8c33d39b17061b220ebc69b656cacb4156bbe"} Oct 14 07:09:36 crc kubenswrapper[5018]: I1014 07:09:36.210559 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:09:36 crc kubenswrapper[5018]: I1014 07:09:36.213683 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="1954e1a7-b433-4ad3-a76a-2147db4b56c1" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" Oct 14 07:09:36 crc kubenswrapper[5018]: I1014 07:09:36.614951 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1954e1a7-b433-4ad3-a76a-2147db4b56c1" path="/var/lib/kubelet/pods/1954e1a7-b433-4ad3-a76a-2147db4b56c1/volumes" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.231289 5018 generic.go:334] "Generic (PLEG): container finished" podID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerID="dab56388c6e4d47b8260023cc1f4193441ec8c3d7619d5bd3c822262d9315d87" exitCode=0 Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.231542 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerDied","Data":"dab56388c6e4d47b8260023cc1f4193441ec8c3d7619d5bd3c822262d9315d87"} Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.508424 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.603993 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604187 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szjgs\" (UniqueName: \"kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604232 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604312 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604355 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604417 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data\") pod \"bfee0da5-22e6-49c5-988c-de2781a8dd53\" (UID: \"bfee0da5-22e6-49c5-988c-de2781a8dd53\") " Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.604745 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.605436 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfee0da5-22e6-49c5-988c-de2781a8dd53-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.609735 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.610800 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs" (OuterVolumeSpecName: "kube-api-access-szjgs") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "kube-api-access-szjgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.617855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts" (OuterVolumeSpecName: "scripts") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.704750 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.708206 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.708235 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.708302 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szjgs\" (UniqueName: \"kubernetes.io/projected/bfee0da5-22e6-49c5-988c-de2781a8dd53-kube-api-access-szjgs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.708314 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.757643 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data" (OuterVolumeSpecName: "config-data") pod "bfee0da5-22e6-49c5-988c-de2781a8dd53" (UID: "bfee0da5-22e6-49c5-988c-de2781a8dd53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:37 crc kubenswrapper[5018]: I1014 07:09:37.811304 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfee0da5-22e6-49c5-988c-de2781a8dd53-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.243235 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfee0da5-22e6-49c5-988c-de2781a8dd53","Type":"ContainerDied","Data":"80098bcadd126150546df569f827c8db720939e4ba0febf6a8b445df434a5ce8"} Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.243285 5018 scope.go:117] "RemoveContainer" containerID="8a7cae35134d5e52028141d136d5b89ec87360e30e3f4a83eb1a443c498199f8" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.243389 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.274930 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.283909 5018 scope.go:117] "RemoveContainer" containerID="dab56388c6e4d47b8260023cc1f4193441ec8c3d7619d5bd3c822262d9315d87" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.284440 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.307898 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:38 crc kubenswrapper[5018]: E1014 07:09:38.308394 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="probe" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.311483 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="probe" Oct 14 07:09:38 crc kubenswrapper[5018]: E1014 07:09:38.311628 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="cinder-scheduler" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.311704 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="cinder-scheduler" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.311917 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="probe" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.311979 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" containerName="cinder-scheduler" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.313961 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.350501 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.353696 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.367168 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.367446 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-central-agent" containerID="cri-o://7fd79b97dcdf57e46266fcbc4e307abb85f811a686b29ba6b8480915b694e9bf" gracePeriod=30 Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.367590 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" containerID="cri-o://6bcb6d325818100d136b12555d42017c0070b89b74b89f2e51e92446c7d8cb3e" gracePeriod=30 Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.367653 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="sg-core" containerID="cri-o://084a9c2a21923f20e85e8d9aca17be12df497b5b8a50604239e63426973d19eb" gracePeriod=30 Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.367681 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-notification-agent" containerID="cri-o://24d86d1fc0be34055cd6378df65e13f2f30420de13c4afe3bea2392b1d5e778a" gracePeriod=30 Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.391680 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.448803 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.448870 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.448940 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgj8\" (UniqueName: \"kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.448966 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.449018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.449037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551039 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551149 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgj8\" (UniqueName: \"kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551184 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551241 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551256 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.551276 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.552317 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.557465 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.559584 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.565402 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.567701 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgj8\" (UniqueName: \"kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.575534 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.636069 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfee0da5-22e6-49c5-988c-de2781a8dd53" path="/var/lib/kubelet/pods/bfee0da5-22e6-49c5-988c-de2781a8dd53/volumes" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.671009 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.955783 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.957205 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.964793 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.965073 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.965087 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 14 07:09:38 crc kubenswrapper[5018]: I1014 07:09:38.969987 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059709 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059755 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059825 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059868 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059910 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059949 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.059970 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpfpk\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.101807 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:09:39 crc kubenswrapper[5018]: W1014 07:09:39.105331 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb738854_dc39_46ff_b965_b69efbbec865.slice/crio-8e44d9b1f49487838ec1b4b24cc4fa83c15f2cce33a69ca620f79e26adfac29e WatchSource:0}: Error finding container 8e44d9b1f49487838ec1b4b24cc4fa83c15f2cce33a69ca620f79e26adfac29e: Status 404 returned error can't find the container with id 8e44d9b1f49487838ec1b4b24cc4fa83c15f2cce33a69ca620f79e26adfac29e Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.161540 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.161635 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162043 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162090 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162112 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162144 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162182 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162203 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpfpk\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.162231 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.166253 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.168336 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.168417 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.171407 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.177598 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.181214 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.188533 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpfpk\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk\") pod \"swift-proxy-65c7ff6d65-cbdkc\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260181 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerID="6bcb6d325818100d136b12555d42017c0070b89b74b89f2e51e92446c7d8cb3e" exitCode=0 Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260212 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerID="084a9c2a21923f20e85e8d9aca17be12df497b5b8a50604239e63426973d19eb" exitCode=2 Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260221 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerID="7fd79b97dcdf57e46266fcbc4e307abb85f811a686b29ba6b8480915b694e9bf" exitCode=0 Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260258 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerDied","Data":"6bcb6d325818100d136b12555d42017c0070b89b74b89f2e51e92446c7d8cb3e"} Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260280 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerDied","Data":"084a9c2a21923f20e85e8d9aca17be12df497b5b8a50604239e63426973d19eb"} Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.260289 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerDied","Data":"7fd79b97dcdf57e46266fcbc4e307abb85f811a686b29ba6b8480915b694e9bf"} Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.269393 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerStarted","Data":"8e44d9b1f49487838ec1b4b24cc4fa83c15f2cce33a69ca620f79e26adfac29e"} Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.282397 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:39 crc kubenswrapper[5018]: I1014 07:09:39.788082 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:09:39 crc kubenswrapper[5018]: W1014 07:09:39.796351 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcced7bf3_2f55_4c00_87f0_b7e17c53ee0e.slice/crio-cf469bc58c9d0775ff8ae9961e65253ba16e2d367ba257c177ff52b2ba6e92d1 WatchSource:0}: Error finding container cf469bc58c9d0775ff8ae9961e65253ba16e2d367ba257c177ff52b2ba6e92d1: Status 404 returned error can't find the container with id cf469bc58c9d0775ff8ae9961e65253ba16e2d367ba257c177ff52b2ba6e92d1 Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.291004 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerStarted","Data":"2b02d899d0cced29508d36139ee74410a8e6281e06ca1fb7f832e68bbfcb9048"} Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.291278 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerStarted","Data":"e00c8d1ad5dfcec3691b831c54fc96d156f290741de76b782f82a372f3292ae3"} Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.291289 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerStarted","Data":"cf469bc58c9d0775ff8ae9961e65253ba16e2d367ba257c177ff52b2ba6e92d1"} Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.291319 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.291336 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.294041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerStarted","Data":"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc"} Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.706002 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 07:09:40 crc kubenswrapper[5018]: I1014 07:09:40.728293 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" podStartSLOduration=2.728273976 podStartE2EDuration="2.728273976s" podCreationTimestamp="2025-10-14 07:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:40.308338769 +0000 UTC m=+1196.892385396" watchObservedRunningTime="2025-10-14 07:09:40.728273976 +0000 UTC m=+1197.312320603" Oct 14 07:09:41 crc kubenswrapper[5018]: I1014 07:09:41.303922 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerStarted","Data":"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f"} Oct 14 07:09:41 crc kubenswrapper[5018]: I1014 07:09:41.323881 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.32386466 podStartE2EDuration="3.32386466s" podCreationTimestamp="2025-10-14 07:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:41.321448331 +0000 UTC m=+1197.905494958" watchObservedRunningTime="2025-10-14 07:09:41.32386466 +0000 UTC m=+1197.907911287" Oct 14 07:09:43 crc kubenswrapper[5018]: I1014 07:09:43.325876 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerID="24d86d1fc0be34055cd6378df65e13f2f30420de13c4afe3bea2392b1d5e778a" exitCode=0 Oct 14 07:09:43 crc kubenswrapper[5018]: I1014 07:09:43.326040 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerDied","Data":"24d86d1fc0be34055cd6378df65e13f2f30420de13c4afe3bea2392b1d5e778a"} Oct 14 07:09:43 crc kubenswrapper[5018]: I1014 07:09:43.671436 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 07:09:46 crc kubenswrapper[5018]: I1014 07:09:46.677704 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.062703 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.118893 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.118942 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.119031 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2lbm\" (UniqueName: \"kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.119067 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.119103 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.119135 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.119160 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml\") pod \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\" (UID: \"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba\") " Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.120256 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.120418 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.127831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts" (OuterVolumeSpecName: "scripts") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.127902 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm" (OuterVolumeSpecName: "kube-api-access-d2lbm") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "kube-api-access-d2lbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.150568 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.200854 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.204818 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data" (OuterVolumeSpecName: "config-data") pod "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" (UID: "2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220811 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2lbm\" (UniqueName: \"kubernetes.io/projected/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-kube-api-access-d2lbm\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220836 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220847 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220910 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220918 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220927 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.220936 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.360961 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba","Type":"ContainerDied","Data":"33f40076df59cfcb0b6d0361033c0236783e33388134a9d4807873db273628a0"} Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.361009 5018 scope.go:117] "RemoveContainer" containerID="6bcb6d325818100d136b12555d42017c0070b89b74b89f2e51e92446c7d8cb3e" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.361114 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.366556 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1719e0be-9beb-4f35-b449-2a90e51ec80c","Type":"ContainerStarted","Data":"672ba654264cd63e079885cb542e177121988e36e45b223e1019d3465e9c3dcc"} Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.381985 5018 scope.go:117] "RemoveContainer" containerID="084a9c2a21923f20e85e8d9aca17be12df497b5b8a50604239e63426973d19eb" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.404085 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.5215117879999998 podStartE2EDuration="13.404070274s" podCreationTimestamp="2025-10-14 07:09:34 +0000 UTC" firstStartedPulling="2025-10-14 07:09:35.773254059 +0000 UTC m=+1192.357300706" lastFinishedPulling="2025-10-14 07:09:46.655812565 +0000 UTC m=+1203.239859192" observedRunningTime="2025-10-14 07:09:47.388982306 +0000 UTC m=+1203.973028933" watchObservedRunningTime="2025-10-14 07:09:47.404070274 +0000 UTC m=+1203.988116901" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.404165 5018 scope.go:117] "RemoveContainer" containerID="24d86d1fc0be34055cd6378df65e13f2f30420de13c4afe3bea2392b1d5e778a" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.404244 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.410792 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.435923 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:47 crc kubenswrapper[5018]: E1014 07:09:47.436325 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-central-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436338 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-central-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: E1014 07:09:47.436349 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="sg-core" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436355 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="sg-core" Oct 14 07:09:47 crc kubenswrapper[5018]: E1014 07:09:47.436375 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-notification-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436381 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-notification-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: E1014 07:09:47.436395 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436401 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436586 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-notification-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436600 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="proxy-httpd" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436633 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="sg-core" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.436647 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" containerName="ceilometer-central-agent" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.438201 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.440715 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.440972 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.442023 5018 scope.go:117] "RemoveContainer" containerID="7fd79b97dcdf57e46266fcbc4e307abb85f811a686b29ba6b8480915b694e9bf" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.471665 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.526808 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.526879 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.526902 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.526934 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.526974 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xpk\" (UniqueName: \"kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.527017 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.527083 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628401 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628475 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628511 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628550 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628570 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628598 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.628648 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xpk\" (UniqueName: \"kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.629106 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.629177 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.631736 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.632586 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.633371 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.634040 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.646841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xpk\" (UniqueName: \"kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk\") pod \"ceilometer-0\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " pod="openstack/ceilometer-0" Oct 14 07:09:47 crc kubenswrapper[5018]: I1014 07:09:47.761914 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:48 crc kubenswrapper[5018]: W1014 07:09:48.250096 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e61489_6388_4870_b906_4006d208bfc1.slice/crio-9d13f097751e87bf29e6228e9bcc2bb29cbcdbce11c0340b4d5b181aa1d32ff2 WatchSource:0}: Error finding container 9d13f097751e87bf29e6228e9bcc2bb29cbcdbce11c0340b4d5b181aa1d32ff2: Status 404 returned error can't find the container with id 9d13f097751e87bf29e6228e9bcc2bb29cbcdbce11c0340b4d5b181aa1d32ff2 Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.258562 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.307766 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.389512 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerStarted","Data":"9d13f097751e87bf29e6228e9bcc2bb29cbcdbce11c0340b4d5b181aa1d32ff2"} Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.621810 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba" path="/var/lib/kubelet/pods/2d4bd3aa-191b-4e22-9ca3-6fd9362fcdba/volumes" Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.702074 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.702286 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-log" containerID="cri-o://e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8" gracePeriod=30 Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.702366 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-httpd" containerID="cri-o://57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7" gracePeriod=30 Oct 14 07:09:48 crc kubenswrapper[5018]: I1014 07:09:48.972191 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.073283 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-k44z4"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.074661 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.087966 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-k44z4"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.156374 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4k94\" (UniqueName: \"kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94\") pod \"nova-api-db-create-k44z4\" (UID: \"fdbc6001-fbcd-4162-b1ab-633b49b15ae9\") " pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.172132 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tcn5l"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.173246 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.194100 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tcn5l"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.258043 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krdzt\" (UniqueName: \"kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt\") pod \"nova-cell0-db-create-tcn5l\" (UID: \"bf484b91-98a3-477a-9476-3c6f52c27297\") " pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.258103 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4k94\" (UniqueName: \"kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94\") pod \"nova-api-db-create-k44z4\" (UID: \"fdbc6001-fbcd-4162-b1ab-633b49b15ae9\") " pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.285098 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5hfgd"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.286158 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.288881 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.289803 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.292320 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4k94\" (UniqueName: \"kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94\") pod \"nova-api-db-create-k44z4\" (UID: \"fdbc6001-fbcd-4162-b1ab-633b49b15ae9\") " pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.334961 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5hfgd"] Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.363281 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df88n\" (UniqueName: \"kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n\") pod \"nova-cell1-db-create-5hfgd\" (UID: \"1a592b94-6c4d-4af2-99a3-860f2629332a\") " pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.363461 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krdzt\" (UniqueName: \"kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt\") pod \"nova-cell0-db-create-tcn5l\" (UID: \"bf484b91-98a3-477a-9476-3c6f52c27297\") " pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.387186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krdzt\" (UniqueName: \"kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt\") pod \"nova-cell0-db-create-tcn5l\" (UID: \"bf484b91-98a3-477a-9476-3c6f52c27297\") " pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.395959 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.405975 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerStarted","Data":"c6634b86a9a180a886c16152724e2e7437c33f0dca9f1ca2095959ea8ce3c76d"} Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.408937 5018 generic.go:334] "Generic (PLEG): container finished" podID="1bf08405-e68a-4eac-9adf-046965311764" containerID="e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8" exitCode=143 Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.409242 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerDied","Data":"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8"} Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.465049 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df88n\" (UniqueName: \"kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n\") pod \"nova-cell1-db-create-5hfgd\" (UID: \"1a592b94-6c4d-4af2-99a3-860f2629332a\") " pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.484736 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df88n\" (UniqueName: \"kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n\") pod \"nova-cell1-db-create-5hfgd\" (UID: \"1a592b94-6c4d-4af2-99a3-860f2629332a\") " pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.493108 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.645268 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:49 crc kubenswrapper[5018]: I1014 07:09:49.816933 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-k44z4"] Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.185109 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tcn5l"] Oct 14 07:09:50 crc kubenswrapper[5018]: W1014 07:09:50.226593 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf484b91_98a3_477a_9476_3c6f52c27297.slice/crio-68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44 WatchSource:0}: Error finding container 68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44: Status 404 returned error can't find the container with id 68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44 Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.308366 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.308636 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-log" containerID="cri-o://448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e" gracePeriod=30 Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.308752 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-httpd" containerID="cri-o://05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c" gracePeriod=30 Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.336195 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5hfgd"] Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.436768 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5hfgd" event={"ID":"1a592b94-6c4d-4af2-99a3-860f2629332a","Type":"ContainerStarted","Data":"f5dede819c4c909d04647bcdb8488f4f9b695916e02d8f61043524545e1ad11e"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.439018 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerStarted","Data":"200ac0bd2b4c8078686c2f5a903a06438ea7f8b63b41e8f951485ea47f9719ba"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.442506 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tcn5l" event={"ID":"bf484b91-98a3-477a-9476-3c6f52c27297","Type":"ContainerStarted","Data":"357f6156d3b1a125fd854af5001fabf852e7fd7d6d4c24cd609e0733b9bb922b"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.442558 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tcn5l" event={"ID":"bf484b91-98a3-477a-9476-3c6f52c27297","Type":"ContainerStarted","Data":"68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.444968 5018 generic.go:334] "Generic (PLEG): container finished" podID="fdbc6001-fbcd-4162-b1ab-633b49b15ae9" containerID="5f58d093c2e1b4df7035a9b6e39bb45dfdc1616b93fda503854192c6a3ac25c4" exitCode=0 Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.445010 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k44z4" event={"ID":"fdbc6001-fbcd-4162-b1ab-633b49b15ae9","Type":"ContainerDied","Data":"5f58d093c2e1b4df7035a9b6e39bb45dfdc1616b93fda503854192c6a3ac25c4"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.445036 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k44z4" event={"ID":"fdbc6001-fbcd-4162-b1ab-633b49b15ae9","Type":"ContainerStarted","Data":"c5aed1329beac4f3ff96e1a6bb8a510ab5091c362adf53435b143d2c88f93953"} Oct 14 07:09:50 crc kubenswrapper[5018]: I1014 07:09:50.469946 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-tcn5l" podStartSLOduration=1.4699234190000001 podStartE2EDuration="1.469923419s" podCreationTimestamp="2025-10-14 07:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:50.455470999 +0000 UTC m=+1207.039517636" watchObservedRunningTime="2025-10-14 07:09:50.469923419 +0000 UTC m=+1207.053970046" Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.454757 5018 generic.go:334] "Generic (PLEG): container finished" podID="1a592b94-6c4d-4af2-99a3-860f2629332a" containerID="91711faedba7fc88d95405468244ef504b36f759b6a39014eb7510244825fb42" exitCode=0 Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.454915 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5hfgd" event={"ID":"1a592b94-6c4d-4af2-99a3-860f2629332a","Type":"ContainerDied","Data":"91711faedba7fc88d95405468244ef504b36f759b6a39014eb7510244825fb42"} Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.460596 5018 generic.go:334] "Generic (PLEG): container finished" podID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerID="448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e" exitCode=143 Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.460848 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerDied","Data":"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e"} Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.463321 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerStarted","Data":"dd62435a405b66d2f8ffc1a1ee1b730ecbb9d1bbe6a9ccc527da1714feb0d72f"} Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.465763 5018 generic.go:334] "Generic (PLEG): container finished" podID="bf484b91-98a3-477a-9476-3c6f52c27297" containerID="357f6156d3b1a125fd854af5001fabf852e7fd7d6d4c24cd609e0733b9bb922b" exitCode=0 Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.465909 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tcn5l" event={"ID":"bf484b91-98a3-477a-9476-3c6f52c27297","Type":"ContainerDied","Data":"357f6156d3b1a125fd854af5001fabf852e7fd7d6d4c24cd609e0733b9bb922b"} Oct 14 07:09:51 crc kubenswrapper[5018]: I1014 07:09:51.864247 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.017326 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4k94\" (UniqueName: \"kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94\") pod \"fdbc6001-fbcd-4162-b1ab-633b49b15ae9\" (UID: \"fdbc6001-fbcd-4162-b1ab-633b49b15ae9\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.028770 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94" (OuterVolumeSpecName: "kube-api-access-p4k94") pod "fdbc6001-fbcd-4162-b1ab-633b49b15ae9" (UID: "fdbc6001-fbcd-4162-b1ab-633b49b15ae9"). InnerVolumeSpecName "kube-api-access-p4k94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.109035 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.119958 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4k94\" (UniqueName: \"kubernetes.io/projected/fdbc6001-fbcd-4162-b1ab-633b49b15ae9-kube-api-access-p4k94\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.311819 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425333 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425683 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425727 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425782 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jgfk\" (UniqueName: \"kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425884 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425932 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425954 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.425987 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts\") pod \"1bf08405-e68a-4eac-9adf-046965311764\" (UID: \"1bf08405-e68a-4eac-9adf-046965311764\") " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.426521 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs" (OuterVolumeSpecName: "logs") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.426678 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.426949 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.426962 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf08405-e68a-4eac-9adf-046965311764-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.429182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.429316 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk" (OuterVolumeSpecName: "kube-api-access-4jgfk") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "kube-api-access-4jgfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.434968 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts" (OuterVolumeSpecName: "scripts") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.455879 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.474441 5018 generic.go:334] "Generic (PLEG): container finished" podID="1bf08405-e68a-4eac-9adf-046965311764" containerID="57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7" exitCode=0 Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.474531 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.474794 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerDied","Data":"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7"} Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.474923 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1bf08405-e68a-4eac-9adf-046965311764","Type":"ContainerDied","Data":"918408d26f8252f98ee0ce66986ce90d4d549d61f6ba6ec79e620053915595c1"} Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.474971 5018 scope.go:117] "RemoveContainer" containerID="57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.476725 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerStarted","Data":"f48e4753327d7ff636bafbcf0c7b6fb93c1d111f316c3fdac17c38da60a4a52c"} Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.476935 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-central-agent" containerID="cri-o://c6634b86a9a180a886c16152724e2e7437c33f0dca9f1ca2095959ea8ce3c76d" gracePeriod=30 Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.477194 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.477230 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="proxy-httpd" containerID="cri-o://f48e4753327d7ff636bafbcf0c7b6fb93c1d111f316c3fdac17c38da60a4a52c" gracePeriod=30 Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.477289 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="sg-core" containerID="cri-o://dd62435a405b66d2f8ffc1a1ee1b730ecbb9d1bbe6a9ccc527da1714feb0d72f" gracePeriod=30 Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.477321 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-notification-agent" containerID="cri-o://200ac0bd2b4c8078686c2f5a903a06438ea7f8b63b41e8f951485ea47f9719ba" gracePeriod=30 Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.489875 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k44z4" event={"ID":"fdbc6001-fbcd-4162-b1ab-633b49b15ae9","Type":"ContainerDied","Data":"c5aed1329beac4f3ff96e1a6bb8a510ab5091c362adf53435b143d2c88f93953"} Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.489921 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5aed1329beac4f3ff96e1a6bb8a510ab5091c362adf53435b143d2c88f93953" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.489930 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k44z4" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.494482 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.501683 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.086670562 podStartE2EDuration="5.501663849s" podCreationTimestamp="2025-10-14 07:09:47 +0000 UTC" firstStartedPulling="2025-10-14 07:09:48.253550589 +0000 UTC m=+1204.837597206" lastFinishedPulling="2025-10-14 07:09:51.668543866 +0000 UTC m=+1208.252590493" observedRunningTime="2025-10-14 07:09:52.497561513 +0000 UTC m=+1209.081608140" watchObservedRunningTime="2025-10-14 07:09:52.501663849 +0000 UTC m=+1209.085710476" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.516721 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data" (OuterVolumeSpecName: "config-data") pod "1bf08405-e68a-4eac-9adf-046965311764" (UID: "1bf08405-e68a-4eac-9adf-046965311764"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.516857 5018 scope.go:117] "RemoveContainer" containerID="e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529130 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529161 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529170 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jgfk\" (UniqueName: \"kubernetes.io/projected/1bf08405-e68a-4eac-9adf-046965311764-kube-api-access-4jgfk\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529199 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529209 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.529217 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf08405-e68a-4eac-9adf-046965311764-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.542232 5018 scope.go:117] "RemoveContainer" containerID="57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7" Oct 14 07:09:52 crc kubenswrapper[5018]: E1014 07:09:52.546744 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7\": container with ID starting with 57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7 not found: ID does not exist" containerID="57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.546799 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7"} err="failed to get container status \"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7\": rpc error: code = NotFound desc = could not find container \"57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7\": container with ID starting with 57be80d7fc80d40781a2d28e0a1f46a3bd11ed347bb7b0faed04d7c7ed61d4c7 not found: ID does not exist" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.546831 5018 scope.go:117] "RemoveContainer" containerID="e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8" Oct 14 07:09:52 crc kubenswrapper[5018]: E1014 07:09:52.547159 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8\": container with ID starting with e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8 not found: ID does not exist" containerID="e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.547211 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8"} err="failed to get container status \"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8\": rpc error: code = NotFound desc = could not find container \"e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8\": container with ID starting with e941094fc485d8285d78764965e8633c292a100b29fa05705296d52f7020cde8 not found: ID does not exist" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.548208 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.630658 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.858698 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.896699 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.908821 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:52 crc kubenswrapper[5018]: E1014 07:09:52.909224 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdbc6001-fbcd-4162-b1ab-633b49b15ae9" containerName="mariadb-database-create" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909236 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdbc6001-fbcd-4162-b1ab-633b49b15ae9" containerName="mariadb-database-create" Oct 14 07:09:52 crc kubenswrapper[5018]: E1014 07:09:52.909265 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-log" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909270 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-log" Oct 14 07:09:52 crc kubenswrapper[5018]: E1014 07:09:52.909278 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-httpd" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909284 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-httpd" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909453 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-httpd" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909466 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdbc6001-fbcd-4162-b1ab-633b49b15ae9" containerName="mariadb-database-create" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.909478 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf08405-e68a-4eac-9adf-046965311764" containerName="glance-log" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.910518 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.914069 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.914257 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:09:52 crc kubenswrapper[5018]: I1014 07:09:52.914789 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051796 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051837 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs5x6\" (UniqueName: \"kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051876 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051912 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051934 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.051982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.052009 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.052045 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.055207 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.056730 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.153423 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krdzt\" (UniqueName: \"kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt\") pod \"bf484b91-98a3-477a-9476-3c6f52c27297\" (UID: \"bf484b91-98a3-477a-9476-3c6f52c27297\") " Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.153690 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df88n\" (UniqueName: \"kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n\") pod \"1a592b94-6c4d-4af2-99a3-860f2629332a\" (UID: \"1a592b94-6c4d-4af2-99a3-860f2629332a\") " Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.153909 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.153943 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.153971 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154018 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154043 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs5x6\" (UniqueName: \"kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154228 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154449 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154589 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.154641 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.155157 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.155380 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.159822 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt" (OuterVolumeSpecName: "kube-api-access-krdzt") pod "bf484b91-98a3-477a-9476-3c6f52c27297" (UID: "bf484b91-98a3-477a-9476-3c6f52c27297"). InnerVolumeSpecName "kube-api-access-krdzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.159905 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.160307 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.161317 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.162105 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.162188 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n" (OuterVolumeSpecName: "kube-api-access-df88n") pod "1a592b94-6c4d-4af2-99a3-860f2629332a" (UID: "1a592b94-6c4d-4af2-99a3-860f2629332a"). InnerVolumeSpecName "kube-api-access-df88n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.181270 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs5x6\" (UniqueName: \"kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.183790 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.256776 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df88n\" (UniqueName: \"kubernetes.io/projected/1a592b94-6c4d-4af2-99a3-860f2629332a-kube-api-access-df88n\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.256820 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krdzt\" (UniqueName: \"kubernetes.io/projected/bf484b91-98a3-477a-9476-3c6f52c27297-kube-api-access-krdzt\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.346726 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.507887 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tcn5l" event={"ID":"bf484b91-98a3-477a-9476-3c6f52c27297","Type":"ContainerDied","Data":"68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44"} Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.508231 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68edd39bb6826fa04bb3361cb983455bfdc9657b07148ceeaca3c756f2ee1c44" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.508342 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tcn5l" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.513098 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5hfgd" event={"ID":"1a592b94-6c4d-4af2-99a3-860f2629332a","Type":"ContainerDied","Data":"f5dede819c4c909d04647bcdb8488f4f9b695916e02d8f61043524545e1ad11e"} Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.513132 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5dede819c4c909d04647bcdb8488f4f9b695916e02d8f61043524545e1ad11e" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.513210 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5hfgd" Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535731 5018 generic.go:334] "Generic (PLEG): container finished" podID="35e61489-6388-4870-b906-4006d208bfc1" containerID="f48e4753327d7ff636bafbcf0c7b6fb93c1d111f316c3fdac17c38da60a4a52c" exitCode=0 Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535760 5018 generic.go:334] "Generic (PLEG): container finished" podID="35e61489-6388-4870-b906-4006d208bfc1" containerID="dd62435a405b66d2f8ffc1a1ee1b730ecbb9d1bbe6a9ccc527da1714feb0d72f" exitCode=2 Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535767 5018 generic.go:334] "Generic (PLEG): container finished" podID="35e61489-6388-4870-b906-4006d208bfc1" containerID="200ac0bd2b4c8078686c2f5a903a06438ea7f8b63b41e8f951485ea47f9719ba" exitCode=0 Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535786 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerDied","Data":"f48e4753327d7ff636bafbcf0c7b6fb93c1d111f316c3fdac17c38da60a4a52c"} Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535817 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerDied","Data":"dd62435a405b66d2f8ffc1a1ee1b730ecbb9d1bbe6a9ccc527da1714feb0d72f"} Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.535828 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerDied","Data":"200ac0bd2b4c8078686c2f5a903a06438ea7f8b63b41e8f951485ea47f9719ba"} Oct 14 07:09:53 crc kubenswrapper[5018]: I1014 07:09:53.924249 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.061181 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185243 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185323 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185353 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgjs8\" (UniqueName: \"kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185388 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185535 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185580 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.185609 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run\") pod \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\" (UID: \"2bd2759b-9c06-4b72-8c05-a3b7617799f8\") " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.186075 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs" (OuterVolumeSpecName: "logs") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.186136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.194830 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts" (OuterVolumeSpecName: "scripts") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.203463 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.203998 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8" (OuterVolumeSpecName: "kube-api-access-tgjs8") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "kube-api-access-tgjs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.231366 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.246148 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.255997 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data" (OuterVolumeSpecName: "config-data") pod "2bd2759b-9c06-4b72-8c05-a3b7617799f8" (UID: "2bd2759b-9c06-4b72-8c05-a3b7617799f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.286961 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.286983 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.286996 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.287006 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.287014 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgjs8\" (UniqueName: \"kubernetes.io/projected/2bd2759b-9c06-4b72-8c05-a3b7617799f8-kube-api-access-tgjs8\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.287025 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd2759b-9c06-4b72-8c05-a3b7617799f8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.287033 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.287041 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd2759b-9c06-4b72-8c05-a3b7617799f8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.307656 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.388433 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.545331 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerStarted","Data":"53f8bbd9b14b8748fd6d68d48207701da31ebb8e7309185ebed668c22c0089ca"} Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.545733 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerStarted","Data":"536abeac8032c0588c7d065bccc2a3860b6b60af148dd561a4d43bf1181ed82f"} Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.547567 5018 generic.go:334] "Generic (PLEG): container finished" podID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerID="05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c" exitCode=0 Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.547609 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerDied","Data":"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c"} Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.547655 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd2759b-9c06-4b72-8c05-a3b7617799f8","Type":"ContainerDied","Data":"d9d718d74dc2da9a17195448db3e42f34a283ac80ee2b0153a12c6c656d6a1fb"} Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.547672 5018 scope.go:117] "RemoveContainer" containerID="05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.547752 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.582922 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.587481 5018 scope.go:117] "RemoveContainer" containerID="448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.590305 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.603703 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.604110 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf484b91-98a3-477a-9476-3c6f52c27297" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604130 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf484b91-98a3-477a-9476-3c6f52c27297" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.604149 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-log" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604156 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-log" Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.604169 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-httpd" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604175 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-httpd" Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.604186 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a592b94-6c4d-4af2-99a3-860f2629332a" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604192 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a592b94-6c4d-4af2-99a3-860f2629332a" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604373 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf484b91-98a3-477a-9476-3c6f52c27297" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604388 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-httpd" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604404 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a592b94-6c4d-4af2-99a3-860f2629332a" containerName="mariadb-database-create" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.604420 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" containerName="glance-log" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.607560 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.645945 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.646406 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.663180 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf08405-e68a-4eac-9adf-046965311764" path="/var/lib/kubelet/pods/1bf08405-e68a-4eac-9adf-046965311764/volumes" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.675692 5018 scope.go:117] "RemoveContainer" containerID="05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c" Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.676154 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c\": container with ID starting with 05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c not found: ID does not exist" containerID="05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.676186 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c"} err="failed to get container status \"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c\": rpc error: code = NotFound desc = could not find container \"05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c\": container with ID starting with 05aa13a40c2faf424de0289241f9f5d05d563c9b182ce18d47509d58bfa2116c not found: ID does not exist" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.676205 5018 scope.go:117] "RemoveContainer" containerID="448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e" Oct 14 07:09:54 crc kubenswrapper[5018]: E1014 07:09:54.677416 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e\": container with ID starting with 448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e not found: ID does not exist" containerID="448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.677442 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e"} err="failed to get container status \"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e\": rpc error: code = NotFound desc = could not find container \"448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e\": container with ID starting with 448438698df5708fe0b2bddf83f9478b6d3f6f8ec1f70a05ea1cfd5a9027b33e not found: ID does not exist" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.683766 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd2759b-9c06-4b72-8c05-a3b7617799f8" path="/var/lib/kubelet/pods/2bd2759b-9c06-4b72-8c05-a3b7617799f8/volumes" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.684513 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695548 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97g5w\" (UniqueName: \"kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695597 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695635 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695665 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695684 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695705 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.695851 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.702799 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.804900 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.804940 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.804986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.805006 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.805030 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.805075 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.805099 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.805209 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97g5w\" (UniqueName: \"kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.808183 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.808272 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.809116 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.811501 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.811664 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.813461 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.820135 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.824109 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97g5w\" (UniqueName: \"kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.843885 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:09:54 crc kubenswrapper[5018]: I1014 07:09:54.965539 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:09:55 crc kubenswrapper[5018]: I1014 07:09:55.520122 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:09:55 crc kubenswrapper[5018]: W1014 07:09:55.520203 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc86ed1d_ae09_4120_8e15_2a960334a853.slice/crio-20b83ce71f8b4a06e3e250e42d4eee3fab91a87919112570e5e5a88ce4c5b716 WatchSource:0}: Error finding container 20b83ce71f8b4a06e3e250e42d4eee3fab91a87919112570e5e5a88ce4c5b716: Status 404 returned error can't find the container with id 20b83ce71f8b4a06e3e250e42d4eee3fab91a87919112570e5e5a88ce4c5b716 Oct 14 07:09:55 crc kubenswrapper[5018]: I1014 07:09:55.561200 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerStarted","Data":"b96379567c4eaad728194e40c42a49fa57be113eff3ccca7f60a2b9c2df42bbb"} Oct 14 07:09:55 crc kubenswrapper[5018]: I1014 07:09:55.566842 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerStarted","Data":"20b83ce71f8b4a06e3e250e42d4eee3fab91a87919112570e5e5a88ce4c5b716"} Oct 14 07:09:55 crc kubenswrapper[5018]: I1014 07:09:55.589605 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.589584891 podStartE2EDuration="3.589584891s" podCreationTimestamp="2025-10-14 07:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:55.584437595 +0000 UTC m=+1212.168484222" watchObservedRunningTime="2025-10-14 07:09:55.589584891 +0000 UTC m=+1212.173631508" Oct 14 07:09:56 crc kubenswrapper[5018]: I1014 07:09:56.577050 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerStarted","Data":"fb571e1ab7e3e5551c801254e6ef0537cf870849ca12640092089f0011af7c29"} Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.588709 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerStarted","Data":"84742c9fbf39092ccf977df6e5a65995900f3bd0bbaf07ff5f7aed100b72b67f"} Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.594505 5018 generic.go:334] "Generic (PLEG): container finished" podID="35e61489-6388-4870-b906-4006d208bfc1" containerID="c6634b86a9a180a886c16152724e2e7437c33f0dca9f1ca2095959ea8ce3c76d" exitCode=0 Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.594558 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerDied","Data":"c6634b86a9a180a886c16152724e2e7437c33f0dca9f1ca2095959ea8ce3c76d"} Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.612835 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.6128110490000003 podStartE2EDuration="3.612811049s" podCreationTimestamp="2025-10-14 07:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:09:57.609815364 +0000 UTC m=+1214.193862011" watchObservedRunningTime="2025-10-14 07:09:57.612811049 +0000 UTC m=+1214.196857726" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.899906 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.967909 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.967999 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968058 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968200 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968242 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52xpk\" (UniqueName: \"kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968308 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968362 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data\") pod \"35e61489-6388-4870-b906-4006d208bfc1\" (UID: \"35e61489-6388-4870-b906-4006d208bfc1\") " Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.968457 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.969070 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.969433 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.974774 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk" (OuterVolumeSpecName: "kube-api-access-52xpk") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "kube-api-access-52xpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:57 crc kubenswrapper[5018]: I1014 07:09:57.990820 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts" (OuterVolumeSpecName: "scripts") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.000531 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.070704 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35e61489-6388-4870-b906-4006d208bfc1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.070740 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.070752 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.070764 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52xpk\" (UniqueName: \"kubernetes.io/projected/35e61489-6388-4870-b906-4006d208bfc1-kube-api-access-52xpk\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.071905 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.085194 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data" (OuterVolumeSpecName: "config-data") pod "35e61489-6388-4870-b906-4006d208bfc1" (UID: "35e61489-6388-4870-b906-4006d208bfc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.172859 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.172921 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e61489-6388-4870-b906-4006d208bfc1-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.605487 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.617982 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35e61489-6388-4870-b906-4006d208bfc1","Type":"ContainerDied","Data":"9d13f097751e87bf29e6228e9bcc2bb29cbcdbce11c0340b4d5b181aa1d32ff2"} Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.618027 5018 scope.go:117] "RemoveContainer" containerID="f48e4753327d7ff636bafbcf0c7b6fb93c1d111f316c3fdac17c38da60a4a52c" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.669567 5018 scope.go:117] "RemoveContainer" containerID="dd62435a405b66d2f8ffc1a1ee1b730ecbb9d1bbe6a9ccc527da1714feb0d72f" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.677561 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.679821 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.707756 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.720836 5018 scope.go:117] "RemoveContainer" containerID="200ac0bd2b4c8078686c2f5a903a06438ea7f8b63b41e8f951485ea47f9719ba" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.723996 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:58 crc kubenswrapper[5018]: E1014 07:09:58.724368 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="sg-core" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724380 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="sg-core" Oct 14 07:09:58 crc kubenswrapper[5018]: E1014 07:09:58.724400 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-central-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724406 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-central-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: E1014 07:09:58.724423 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-notification-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724429 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-notification-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: E1014 07:09:58.724452 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="proxy-httpd" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724457 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="proxy-httpd" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724645 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="proxy-httpd" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724660 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-central-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724677 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="ceilometer-notification-agent" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.724688 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e61489-6388-4870-b906-4006d208bfc1" containerName="sg-core" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.726242 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.732960 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.733172 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.742540 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.783014 5018 scope.go:117] "RemoveContainer" containerID="c6634b86a9a180a886c16152724e2e7437c33f0dca9f1ca2095959ea8ce3c76d" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.784072 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.784916 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bc7b47c6d-sbkf8" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-api" containerID="cri-o://cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab" gracePeriod=30 Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.784553 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bc7b47c6d-sbkf8" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-httpd" containerID="cri-o://129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40" gracePeriod=30 Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.819988 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820107 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfrk5\" (UniqueName: \"kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820177 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820270 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820328 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820344 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.820365 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922236 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfrk5\" (UniqueName: \"kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922550 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922754 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922829 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922868 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922907 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.922995 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.923329 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.923514 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.936499 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.937444 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.938211 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.954263 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:58 crc kubenswrapper[5018]: I1014 07:09:58.978541 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfrk5\" (UniqueName: \"kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5\") pod \"ceilometer-0\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " pod="openstack/ceilometer-0" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.043114 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.240984 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f4bd-account-create-2g595"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.242138 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.247990 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.251700 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f4bd-account-create-2g595"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.331591 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5vq4\" (UniqueName: \"kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4\") pod \"nova-api-f4bd-account-create-2g595\" (UID: \"b736e7b0-e21d-4859-8c8f-033c5a62e7c8\") " pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.433764 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5vq4\" (UniqueName: \"kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4\") pod \"nova-api-f4bd-account-create-2g595\" (UID: \"b736e7b0-e21d-4859-8c8f-033c5a62e7c8\") " pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.447985 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-fab3-account-create-v5ss6"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.449146 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.454522 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.464578 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fab3-account-create-v5ss6"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.467044 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5vq4\" (UniqueName: \"kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4\") pod \"nova-api-f4bd-account-create-2g595\" (UID: \"b736e7b0-e21d-4859-8c8f-033c5a62e7c8\") " pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.535779 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfhzb\" (UniqueName: \"kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb\") pod \"nova-cell0-fab3-account-create-v5ss6\" (UID: \"77f4e13a-f72b-46bb-92d0-32eaaf87c46b\") " pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.549434 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.565997 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.616312 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerStarted","Data":"d855ea76ed0c9fb82dd238311c02fe8c5898d27048ffa9ca3b24846e88441df2"} Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.618416 5018 generic.go:334] "Generic (PLEG): container finished" podID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerID="129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40" exitCode=0 Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.618491 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerDied","Data":"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40"} Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.634167 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-55b4-account-create-72fds"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.635261 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.638159 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfhzb\" (UniqueName: \"kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb\") pod \"nova-cell0-fab3-account-create-v5ss6\" (UID: \"77f4e13a-f72b-46bb-92d0-32eaaf87c46b\") " pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.638752 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.643476 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-55b4-account-create-72fds"] Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.667504 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfhzb\" (UniqueName: \"kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb\") pod \"nova-cell0-fab3-account-create-v5ss6\" (UID: \"77f4e13a-f72b-46bb-92d0-32eaaf87c46b\") " pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.739654 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhfkq\" (UniqueName: \"kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq\") pod \"nova-cell1-55b4-account-create-72fds\" (UID: \"0723b8b5-790c-4dc7-874f-4e7456b8cfe8\") " pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.782047 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.841713 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhfkq\" (UniqueName: \"kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq\") pod \"nova-cell1-55b4-account-create-72fds\" (UID: \"0723b8b5-790c-4dc7-874f-4e7456b8cfe8\") " pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:09:59 crc kubenswrapper[5018]: I1014 07:09:59.861452 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhfkq\" (UniqueName: \"kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq\") pod \"nova-cell1-55b4-account-create-72fds\" (UID: \"0723b8b5-790c-4dc7-874f-4e7456b8cfe8\") " pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.030024 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.031383 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f4bd-account-create-2g595"] Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.285013 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fab3-account-create-v5ss6"] Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.576300 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-55b4-account-create-72fds"] Oct 14 07:10:00 crc kubenswrapper[5018]: W1014 07:10:00.594318 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0723b8b5_790c_4dc7_874f_4e7456b8cfe8.slice/crio-9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a WatchSource:0}: Error finding container 9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a: Status 404 returned error can't find the container with id 9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.622139 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e61489-6388-4870-b906-4006d208bfc1" path="/var/lib/kubelet/pods/35e61489-6388-4870-b906-4006d208bfc1/volumes" Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.641285 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-55b4-account-create-72fds" event={"ID":"0723b8b5-790c-4dc7-874f-4e7456b8cfe8","Type":"ContainerStarted","Data":"9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a"} Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.643034 5018 generic.go:334] "Generic (PLEG): container finished" podID="b736e7b0-e21d-4859-8c8f-033c5a62e7c8" containerID="9903204a108ca41d3c2402336f8b8ea88b9a08b94770a7155c078e2492dacc43" exitCode=0 Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.644192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f4bd-account-create-2g595" event={"ID":"b736e7b0-e21d-4859-8c8f-033c5a62e7c8","Type":"ContainerDied","Data":"9903204a108ca41d3c2402336f8b8ea88b9a08b94770a7155c078e2492dacc43"} Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.644216 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f4bd-account-create-2g595" event={"ID":"b736e7b0-e21d-4859-8c8f-033c5a62e7c8","Type":"ContainerStarted","Data":"6c9f6d770581da9cf22fa4bc34cf8dbeb50eaed50c2ecf552bff963839b10864"} Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.645694 5018 generic.go:334] "Generic (PLEG): container finished" podID="77f4e13a-f72b-46bb-92d0-32eaaf87c46b" containerID="44ea1af93d522ee762df6bcf89d1c6d4e233fb449c6c50287d54ddf3ebb98aac" exitCode=0 Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.645747 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fab3-account-create-v5ss6" event={"ID":"77f4e13a-f72b-46bb-92d0-32eaaf87c46b","Type":"ContainerDied","Data":"44ea1af93d522ee762df6bcf89d1c6d4e233fb449c6c50287d54ddf3ebb98aac"} Oct 14 07:10:00 crc kubenswrapper[5018]: I1014 07:10:00.645778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fab3-account-create-v5ss6" event={"ID":"77f4e13a-f72b-46bb-92d0-32eaaf87c46b","Type":"ContainerStarted","Data":"41a41b245762fe78df8668e970e73ff18d4e7f530196d3b3ff6537191b8f04e0"} Oct 14 07:10:01 crc kubenswrapper[5018]: I1014 07:10:01.658916 5018 generic.go:334] "Generic (PLEG): container finished" podID="0723b8b5-790c-4dc7-874f-4e7456b8cfe8" containerID="eafbe1d749e70aac5a62d1ff29e85ed71cd319d9a0ecfadcdd4aec4654df7467" exitCode=0 Oct 14 07:10:01 crc kubenswrapper[5018]: I1014 07:10:01.659221 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-55b4-account-create-72fds" event={"ID":"0723b8b5-790c-4dc7-874f-4e7456b8cfe8","Type":"ContainerDied","Data":"eafbe1d749e70aac5a62d1ff29e85ed71cd319d9a0ecfadcdd4aec4654df7467"} Oct 14 07:10:01 crc kubenswrapper[5018]: I1014 07:10:01.662877 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerStarted","Data":"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5"} Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.156326 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.162043 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.325023 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5vq4\" (UniqueName: \"kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4\") pod \"b736e7b0-e21d-4859-8c8f-033c5a62e7c8\" (UID: \"b736e7b0-e21d-4859-8c8f-033c5a62e7c8\") " Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.325166 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfhzb\" (UniqueName: \"kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb\") pod \"77f4e13a-f72b-46bb-92d0-32eaaf87c46b\" (UID: \"77f4e13a-f72b-46bb-92d0-32eaaf87c46b\") " Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.329158 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4" (OuterVolumeSpecName: "kube-api-access-p5vq4") pod "b736e7b0-e21d-4859-8c8f-033c5a62e7c8" (UID: "b736e7b0-e21d-4859-8c8f-033c5a62e7c8"). InnerVolumeSpecName "kube-api-access-p5vq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.330013 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb" (OuterVolumeSpecName: "kube-api-access-vfhzb") pod "77f4e13a-f72b-46bb-92d0-32eaaf87c46b" (UID: "77f4e13a-f72b-46bb-92d0-32eaaf87c46b"). InnerVolumeSpecName "kube-api-access-vfhzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.427529 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5vq4\" (UniqueName: \"kubernetes.io/projected/b736e7b0-e21d-4859-8c8f-033c5a62e7c8-kube-api-access-p5vq4\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.427858 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfhzb\" (UniqueName: \"kubernetes.io/projected/77f4e13a-f72b-46bb-92d0-32eaaf87c46b-kube-api-access-vfhzb\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.672969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerStarted","Data":"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405"} Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.673267 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerStarted","Data":"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb"} Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.674216 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fab3-account-create-v5ss6" event={"ID":"77f4e13a-f72b-46bb-92d0-32eaaf87c46b","Type":"ContainerDied","Data":"41a41b245762fe78df8668e970e73ff18d4e7f530196d3b3ff6537191b8f04e0"} Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.674256 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41a41b245762fe78df8668e970e73ff18d4e7f530196d3b3ff6537191b8f04e0" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.674237 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fab3-account-create-v5ss6" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.678961 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f4bd-account-create-2g595" Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.679868 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f4bd-account-create-2g595" event={"ID":"b736e7b0-e21d-4859-8c8f-033c5a62e7c8","Type":"ContainerDied","Data":"6c9f6d770581da9cf22fa4bc34cf8dbeb50eaed50c2ecf552bff963839b10864"} Oct 14 07:10:02 crc kubenswrapper[5018]: I1014 07:10:02.679897 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c9f6d770581da9cf22fa4bc34cf8dbeb50eaed50c2ecf552bff963839b10864" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.010481 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.139400 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhfkq\" (UniqueName: \"kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq\") pod \"0723b8b5-790c-4dc7-874f-4e7456b8cfe8\" (UID: \"0723b8b5-790c-4dc7-874f-4e7456b8cfe8\") " Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.143845 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq" (OuterVolumeSpecName: "kube-api-access-lhfkq") pod "0723b8b5-790c-4dc7-874f-4e7456b8cfe8" (UID: "0723b8b5-790c-4dc7-874f-4e7456b8cfe8"). InnerVolumeSpecName "kube-api-access-lhfkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.241895 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhfkq\" (UniqueName: \"kubernetes.io/projected/0723b8b5-790c-4dc7-874f-4e7456b8cfe8-kube-api-access-lhfkq\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.347433 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.347829 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.383905 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.393024 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.688242 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-55b4-account-create-72fds" event={"ID":"0723b8b5-790c-4dc7-874f-4e7456b8cfe8","Type":"ContainerDied","Data":"9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a"} Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.688284 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9365794bb95e2204fcae429ba5b68ba6121b8f5c803f3dd55f9d7a04d26a314a" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.688291 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-55b4-account-create-72fds" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.688651 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:10:03 crc kubenswrapper[5018]: I1014 07:10:03.688684 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.572304 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-skccn"] Oct 14 07:10:04 crc kubenswrapper[5018]: E1014 07:10:04.573082 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f4e13a-f72b-46bb-92d0-32eaaf87c46b" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573106 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f4e13a-f72b-46bb-92d0-32eaaf87c46b" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: E1014 07:10:04.573133 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0723b8b5-790c-4dc7-874f-4e7456b8cfe8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573143 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0723b8b5-790c-4dc7-874f-4e7456b8cfe8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: E1014 07:10:04.573176 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b736e7b0-e21d-4859-8c8f-033c5a62e7c8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573187 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b736e7b0-e21d-4859-8c8f-033c5a62e7c8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573407 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f4e13a-f72b-46bb-92d0-32eaaf87c46b" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573436 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b736e7b0-e21d-4859-8c8f-033c5a62e7c8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.573456 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0723b8b5-790c-4dc7-874f-4e7456b8cfe8" containerName="mariadb-account-create" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.574265 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.576033 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.576125 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tnjd5" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.576388 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.633445 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-skccn"] Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.665740 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.665954 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngmfl\" (UniqueName: \"kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.666059 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.666134 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.700435 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerStarted","Data":"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01"} Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.700803 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.732072 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.811809057 podStartE2EDuration="6.732053989s" podCreationTimestamp="2025-10-14 07:09:58 +0000 UTC" firstStartedPulling="2025-10-14 07:09:59.563478652 +0000 UTC m=+1216.147525269" lastFinishedPulling="2025-10-14 07:10:03.483723574 +0000 UTC m=+1220.067770201" observedRunningTime="2025-10-14 07:10:04.721795168 +0000 UTC m=+1221.305841795" watchObservedRunningTime="2025-10-14 07:10:04.732053989 +0000 UTC m=+1221.316100616" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.767983 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.768087 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngmfl\" (UniqueName: \"kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.768124 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.768153 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.774595 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.791643 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.791727 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngmfl\" (UniqueName: \"kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.792754 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data\") pod \"nova-cell0-conductor-db-sync-skccn\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.896476 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.965781 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:04 crc kubenswrapper[5018]: I1014 07:10:04.965816 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.028365 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.053156 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.389898 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-skccn"] Oct 14 07:10:05 crc kubenswrapper[5018]: W1014 07:10:05.399888 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14458a1a_5f03_4fff_8983_1547756eb34c.slice/crio-8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6 WatchSource:0}: Error finding container 8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6: Status 404 returned error can't find the container with id 8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6 Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.710026 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-skccn" event={"ID":"14458a1a-5f03-4fff-8983-1547756eb34c","Type":"ContainerStarted","Data":"8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6"} Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.711132 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.711158 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.923767 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.924094 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:10:05 crc kubenswrapper[5018]: I1014 07:10:05.925401 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.449863 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.631047 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config\") pod \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.631125 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config\") pod \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.631518 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs\") pod \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.631563 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle\") pod \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.631596 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6q8r\" (UniqueName: \"kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r\") pod \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\" (UID: \"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9\") " Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.645008 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r" (OuterVolumeSpecName: "kube-api-access-f6q8r") pod "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" (UID: "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9"). InnerVolumeSpecName "kube-api-access-f6q8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.651497 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" (UID: "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.702138 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config" (OuterVolumeSpecName: "config") pod "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" (UID: "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.714505 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" (UID: "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.736822 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.736849 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.736859 5018 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.736868 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6q8r\" (UniqueName: \"kubernetes.io/projected/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-kube-api-access-f6q8r\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.738912 5018 generic.go:334] "Generic (PLEG): container finished" podID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerID="cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab" exitCode=0 Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.738970 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bc7b47c6d-sbkf8" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.738991 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerDied","Data":"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab"} Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.739018 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bc7b47c6d-sbkf8" event={"ID":"2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9","Type":"ContainerDied","Data":"b3a2e763c70f090a60ded9af125db1d8a8dea1208209fd50ed39ec58f8f5cbd9"} Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.739033 5018 scope.go:117] "RemoveContainer" containerID="129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.738981 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.739717 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.767399 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" (UID: "2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.770024 5018 scope.go:117] "RemoveContainer" containerID="cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.790160 5018 scope.go:117] "RemoveContainer" containerID="129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40" Oct 14 07:10:07 crc kubenswrapper[5018]: E1014 07:10:07.790537 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40\": container with ID starting with 129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40 not found: ID does not exist" containerID="129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.790564 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40"} err="failed to get container status \"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40\": rpc error: code = NotFound desc = could not find container \"129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40\": container with ID starting with 129c35a200ad7c46ea84daafc453455537d7d3e7a657e39b94643446378ace40 not found: ID does not exist" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.790583 5018 scope.go:117] "RemoveContainer" containerID="cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab" Oct 14 07:10:07 crc kubenswrapper[5018]: E1014 07:10:07.790891 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab\": container with ID starting with cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab not found: ID does not exist" containerID="cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.790912 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab"} err="failed to get container status \"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab\": rpc error: code = NotFound desc = could not find container \"cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab\": container with ID starting with cc45e63776c5d07f73ed649dea2f5628c563e1a7e95e1057f579fcbfc58986ab not found: ID does not exist" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.838153 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.866471 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:07 crc kubenswrapper[5018]: I1014 07:10:07.876379 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.068840 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.077534 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bc7b47c6d-sbkf8"] Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.593859 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.594092 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-central-agent" containerID="cri-o://5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" gracePeriod=30 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.594183 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="proxy-httpd" containerID="cri-o://7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" gracePeriod=30 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.594206 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-notification-agent" containerID="cri-o://359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" gracePeriod=30 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.594175 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="sg-core" containerID="cri-o://8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" gracePeriod=30 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.628398 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" path="/var/lib/kubelet/pods/2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9/volumes" Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.773211 5018 generic.go:334] "Generic (PLEG): container finished" podID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerID="7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" exitCode=0 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.773263 5018 generic.go:334] "Generic (PLEG): container finished" podID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerID="8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" exitCode=2 Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.773347 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerDied","Data":"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01"} Oct 14 07:10:08 crc kubenswrapper[5018]: I1014 07:10:08.773375 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerDied","Data":"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb"} Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.348039 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476506 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfrk5\" (UniqueName: \"kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476568 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476604 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476723 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476830 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476853 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.476867 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml\") pod \"bb21ffbf-57f6-4581-8407-387977c3b2e4\" (UID: \"bb21ffbf-57f6-4581-8407-387977c3b2e4\") " Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.488925 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.491214 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.492091 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts" (OuterVolumeSpecName: "scripts") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.499810 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5" (OuterVolumeSpecName: "kube-api-access-wfrk5") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "kube-api-access-wfrk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.531826 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.578354 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.578392 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb21ffbf-57f6-4581-8407-387977c3b2e4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.578404 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.578416 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfrk5\" (UniqueName: \"kubernetes.io/projected/bb21ffbf-57f6-4581-8407-387977c3b2e4-kube-api-access-wfrk5\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.578426 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.619081 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data" (OuterVolumeSpecName: "config-data") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.626194 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb21ffbf-57f6-4581-8407-387977c3b2e4" (UID: "bb21ffbf-57f6-4581-8407-387977c3b2e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.680484 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.680518 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb21ffbf-57f6-4581-8407-387977c3b2e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.793920 5018 generic.go:334] "Generic (PLEG): container finished" podID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerID="359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" exitCode=0 Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794264 5018 generic.go:334] "Generic (PLEG): container finished" podID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerID="5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" exitCode=0 Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794026 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerDied","Data":"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405"} Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794404 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerDied","Data":"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5"} Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794418 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb21ffbf-57f6-4581-8407-387977c3b2e4","Type":"ContainerDied","Data":"d855ea76ed0c9fb82dd238311c02fe8c5898d27048ffa9ca3b24846e88441df2"} Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794434 5018 scope.go:117] "RemoveContainer" containerID="7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.794003 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.836554 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.848638 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.858943 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859317 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-api" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859328 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-api" Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859362 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="proxy-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859368 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="proxy-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859380 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-notification-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859386 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-notification-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859401 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859407 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859417 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-central-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859423 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-central-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: E1014 07:10:09.859437 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="sg-core" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.859442 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="sg-core" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860309 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="proxy-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860325 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-api" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860343 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf4b6a8-03bb-4e2b-a59f-73d3a98442a9" containerName="neutron-httpd" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860354 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-notification-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860363 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="ceilometer-central-agent" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.860376 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" containerName="sg-core" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.861980 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.868011 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.868346 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.873007 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.986765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.986817 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.986890 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtc4k\" (UniqueName: \"kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.986920 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.986936 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.987046 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:09 crc kubenswrapper[5018]: I1014 07:10:09.987116 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.088950 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.089642 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.089850 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.090050 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtc4k\" (UniqueName: \"kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.090501 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.090528 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.090606 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.091067 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.090970 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.093531 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.094675 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.094819 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.095392 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.106223 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtc4k\" (UniqueName: \"kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k\") pod \"ceilometer-0\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.185531 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:10 crc kubenswrapper[5018]: I1014 07:10:10.618788 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb21ffbf-57f6-4581-8407-387977c3b2e4" path="/var/lib/kubelet/pods/bb21ffbf-57f6-4581-8407-387977c3b2e4/volumes" Oct 14 07:10:14 crc kubenswrapper[5018]: I1014 07:10:14.756267 5018 scope.go:117] "RemoveContainer" containerID="8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" Oct 14 07:10:14 crc kubenswrapper[5018]: I1014 07:10:14.778745 5018 scope.go:117] "RemoveContainer" containerID="359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.101102 5018 scope.go:117] "RemoveContainer" containerID="5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.135515 5018 scope.go:117] "RemoveContainer" containerID="7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" Oct 14 07:10:15 crc kubenswrapper[5018]: E1014 07:10:15.135986 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01\": container with ID starting with 7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01 not found: ID does not exist" containerID="7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136026 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01"} err="failed to get container status \"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01\": rpc error: code = NotFound desc = could not find container \"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01\": container with ID starting with 7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136054 5018 scope.go:117] "RemoveContainer" containerID="8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" Oct 14 07:10:15 crc kubenswrapper[5018]: E1014 07:10:15.136359 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb\": container with ID starting with 8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb not found: ID does not exist" containerID="8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136381 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb"} err="failed to get container status \"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb\": rpc error: code = NotFound desc = could not find container \"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb\": container with ID starting with 8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136393 5018 scope.go:117] "RemoveContainer" containerID="359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" Oct 14 07:10:15 crc kubenswrapper[5018]: E1014 07:10:15.136659 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405\": container with ID starting with 359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405 not found: ID does not exist" containerID="359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136677 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405"} err="failed to get container status \"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405\": rpc error: code = NotFound desc = could not find container \"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405\": container with ID starting with 359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.136688 5018 scope.go:117] "RemoveContainer" containerID="5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" Oct 14 07:10:15 crc kubenswrapper[5018]: E1014 07:10:15.137132 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5\": container with ID starting with 5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5 not found: ID does not exist" containerID="5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137172 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5"} err="failed to get container status \"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5\": rpc error: code = NotFound desc = could not find container \"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5\": container with ID starting with 5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137199 5018 scope.go:117] "RemoveContainer" containerID="7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137497 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01"} err="failed to get container status \"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01\": rpc error: code = NotFound desc = could not find container \"7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01\": container with ID starting with 7c6168e09cede8c535798cf1bcd055b82d1139ae18622a53e0cd7c4dbd033d01 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137520 5018 scope.go:117] "RemoveContainer" containerID="8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137788 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb"} err="failed to get container status \"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb\": rpc error: code = NotFound desc = could not find container \"8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb\": container with ID starting with 8bd13922d405913e311aceb2a510cc7fbdef5dc0624b96ccf2a797fbb39074bb not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.137807 5018 scope.go:117] "RemoveContainer" containerID="359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.138043 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405"} err="failed to get container status \"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405\": rpc error: code = NotFound desc = could not find container \"359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405\": container with ID starting with 359717b41cca30b578d5675f8fa89229a2a908acf0717801e8bdd8115f65c405 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.138064 5018 scope.go:117] "RemoveContainer" containerID="5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.138252 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5"} err="failed to get container status \"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5\": rpc error: code = NotFound desc = could not find container \"5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5\": container with ID starting with 5e7b5e5b4ea05fad5b2e957d0d0a8f3987224a2d45ac645818ec76f010a6d1e5 not found: ID does not exist" Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.227790 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.247313 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.884061 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerStarted","Data":"51a765d4c50444ea679f768356bb8d41951693ab9e02155b226341fa0f30d6ef"} Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.888612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-skccn" event={"ID":"14458a1a-5f03-4fff-8983-1547756eb34c","Type":"ContainerStarted","Data":"5deaacab0ff8da05b088f6bd2538d21f0764eafa9e3a35c45d78906b2a2b94c7"} Oct 14 07:10:15 crc kubenswrapper[5018]: I1014 07:10:15.912071 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-skccn" podStartSLOduration=2.467625389 podStartE2EDuration="11.912050322s" podCreationTimestamp="2025-10-14 07:10:04 +0000 UTC" firstStartedPulling="2025-10-14 07:10:05.401474204 +0000 UTC m=+1221.985520831" lastFinishedPulling="2025-10-14 07:10:14.845899127 +0000 UTC m=+1231.429945764" observedRunningTime="2025-10-14 07:10:15.908549442 +0000 UTC m=+1232.492596079" watchObservedRunningTime="2025-10-14 07:10:15.912050322 +0000 UTC m=+1232.496096949" Oct 14 07:10:16 crc kubenswrapper[5018]: I1014 07:10:16.903675 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerStarted","Data":"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699"} Oct 14 07:10:17 crc kubenswrapper[5018]: I1014 07:10:17.915102 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerStarted","Data":"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44"} Oct 14 07:10:18 crc kubenswrapper[5018]: I1014 07:10:18.924995 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerStarted","Data":"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf"} Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.939341 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerStarted","Data":"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d"} Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.939714 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-central-agent" containerID="cri-o://b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699" gracePeriod=30 Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.940002 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.940605 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="sg-core" containerID="cri-o://6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf" gracePeriod=30 Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.940781 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="proxy-httpd" containerID="cri-o://5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d" gracePeriod=30 Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.940915 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-notification-agent" containerID="cri-o://85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44" gracePeriod=30 Oct 14 07:10:19 crc kubenswrapper[5018]: I1014 07:10:19.976579 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.632603105 podStartE2EDuration="10.976562341s" podCreationTimestamp="2025-10-14 07:10:09 +0000 UTC" firstStartedPulling="2025-10-14 07:10:15.250068237 +0000 UTC m=+1231.834114874" lastFinishedPulling="2025-10-14 07:10:19.594027493 +0000 UTC m=+1236.178074110" observedRunningTime="2025-10-14 07:10:19.965830997 +0000 UTC m=+1236.549877634" watchObservedRunningTime="2025-10-14 07:10:19.976562341 +0000 UTC m=+1236.560608968" Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971153 5018 generic.go:334] "Generic (PLEG): container finished" podID="3014d924-866b-4e16-ba70-8fe7664747fd" containerID="5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d" exitCode=0 Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971446 5018 generic.go:334] "Generic (PLEG): container finished" podID="3014d924-866b-4e16-ba70-8fe7664747fd" containerID="6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf" exitCode=2 Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971459 5018 generic.go:334] "Generic (PLEG): container finished" podID="3014d924-866b-4e16-ba70-8fe7664747fd" containerID="85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44" exitCode=0 Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971253 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerDied","Data":"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d"} Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971499 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerDied","Data":"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf"} Oct 14 07:10:20 crc kubenswrapper[5018]: I1014 07:10:20.971516 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerDied","Data":"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44"} Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.686026 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824050 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824094 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824139 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824193 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824219 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824470 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824668 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824266 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.824960 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtc4k\" (UniqueName: \"kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k\") pod \"3014d924-866b-4e16-ba70-8fe7664747fd\" (UID: \"3014d924-866b-4e16-ba70-8fe7664747fd\") " Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.825410 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.825430 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3014d924-866b-4e16-ba70-8fe7664747fd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.834900 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k" (OuterVolumeSpecName: "kube-api-access-mtc4k") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "kube-api-access-mtc4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.836339 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts" (OuterVolumeSpecName: "scripts") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.856332 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.915108 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.927530 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.927556 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.927565 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtc4k\" (UniqueName: \"kubernetes.io/projected/3014d924-866b-4e16-ba70-8fe7664747fd-kube-api-access-mtc4k\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.927575 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.940651 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data" (OuterVolumeSpecName: "config-data") pod "3014d924-866b-4e16-ba70-8fe7664747fd" (UID: "3014d924-866b-4e16-ba70-8fe7664747fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.993751 5018 generic.go:334] "Generic (PLEG): container finished" podID="3014d924-866b-4e16-ba70-8fe7664747fd" containerID="b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699" exitCode=0 Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.993815 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerDied","Data":"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699"} Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.993833 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.993849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3014d924-866b-4e16-ba70-8fe7664747fd","Type":"ContainerDied","Data":"51a765d4c50444ea679f768356bb8d41951693ab9e02155b226341fa0f30d6ef"} Oct 14 07:10:22 crc kubenswrapper[5018]: I1014 07:10:22.993871 5018 scope.go:117] "RemoveContainer" containerID="5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.029450 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3014d924-866b-4e16-ba70-8fe7664747fd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.039794 5018 scope.go:117] "RemoveContainer" containerID="6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.039946 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.067672 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.080003 5018 scope.go:117] "RemoveContainer" containerID="85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.097901 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.098555 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-notification-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.098579 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-notification-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.098608 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="proxy-httpd" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.098642 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="proxy-httpd" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.098658 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="sg-core" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.098666 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="sg-core" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.098767 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-central-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.098788 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-central-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.099106 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="proxy-httpd" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.099139 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-notification-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.099148 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="sg-core" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.099163 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" containerName="ceilometer-central-agent" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.101342 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.105981 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.106189 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.111690 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.116826 5018 scope.go:117] "RemoveContainer" containerID="b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.194290 5018 scope.go:117] "RemoveContainer" containerID="5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.194911 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d\": container with ID starting with 5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d not found: ID does not exist" containerID="5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.194947 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d"} err="failed to get container status \"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d\": rpc error: code = NotFound desc = could not find container \"5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d\": container with ID starting with 5b082772eb43924eabc6e8776cf468fa24f48f1ab1ce85ce4bf944e438b9336d not found: ID does not exist" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.194966 5018 scope.go:117] "RemoveContainer" containerID="6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.198850 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf\": container with ID starting with 6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf not found: ID does not exist" containerID="6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.198884 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf"} err="failed to get container status \"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf\": rpc error: code = NotFound desc = could not find container \"6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf\": container with ID starting with 6d517cc9f7ff70573b6ea6774b0762ff58486055e345b9d4d37c374b96096fdf not found: ID does not exist" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.198904 5018 scope.go:117] "RemoveContainer" containerID="85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.202914 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44\": container with ID starting with 85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44 not found: ID does not exist" containerID="85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.202946 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44"} err="failed to get container status \"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44\": rpc error: code = NotFound desc = could not find container \"85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44\": container with ID starting with 85ba4344f728dbd33789ae9d9cf853a83549c4d5ad770d5d3065d249965f1e44 not found: ID does not exist" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.202984 5018 scope.go:117] "RemoveContainer" containerID="b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699" Oct 14 07:10:23 crc kubenswrapper[5018]: E1014 07:10:23.206899 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699\": container with ID starting with b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699 not found: ID does not exist" containerID="b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.206944 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699"} err="failed to get container status \"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699\": rpc error: code = NotFound desc = could not find container \"b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699\": container with ID starting with b6b331d92aa243498263e37468f3a4cc5917a8c6d9a7b0904c7edd623d7dd699 not found: ID does not exist" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.232700 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gntcn\" (UniqueName: \"kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.232739 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.232838 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.233026 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.233185 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.233257 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.233314 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335339 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335434 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335486 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335519 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335545 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335584 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335601 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gntcn\" (UniqueName: \"kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.335864 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.336927 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.340280 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.341054 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.342043 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.342141 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.356297 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gntcn\" (UniqueName: \"kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn\") pod \"ceilometer-0\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.454947 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:10:23 crc kubenswrapper[5018]: W1014 07:10:23.977086 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37976e65_9d86_4384_adcc_8e4f0cee52c7.slice/crio-e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27 WatchSource:0}: Error finding container e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27: Status 404 returned error can't find the container with id e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27 Oct 14 07:10:23 crc kubenswrapper[5018]: I1014 07:10:23.981510 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:24 crc kubenswrapper[5018]: I1014 07:10:24.006157 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerStarted","Data":"e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27"} Oct 14 07:10:24 crc kubenswrapper[5018]: I1014 07:10:24.615420 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3014d924-866b-4e16-ba70-8fe7664747fd" path="/var/lib/kubelet/pods/3014d924-866b-4e16-ba70-8fe7664747fd/volumes" Oct 14 07:10:25 crc kubenswrapper[5018]: I1014 07:10:25.032361 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerStarted","Data":"78d244c57eec48f01d053b75b6b15df44756f1a86d158f3abd8d35b740e667d8"} Oct 14 07:10:26 crc kubenswrapper[5018]: I1014 07:10:26.042080 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerStarted","Data":"0b41efaf38effb59e9892333dafa31b2432c64f40e8aa04ffd1d117c8bf57289"} Oct 14 07:10:26 crc kubenswrapper[5018]: I1014 07:10:26.044378 5018 generic.go:334] "Generic (PLEG): container finished" podID="14458a1a-5f03-4fff-8983-1547756eb34c" containerID="5deaacab0ff8da05b088f6bd2538d21f0764eafa9e3a35c45d78906b2a2b94c7" exitCode=0 Oct 14 07:10:26 crc kubenswrapper[5018]: I1014 07:10:26.044432 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-skccn" event={"ID":"14458a1a-5f03-4fff-8983-1547756eb34c","Type":"ContainerDied","Data":"5deaacab0ff8da05b088f6bd2538d21f0764eafa9e3a35c45d78906b2a2b94c7"} Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.060417 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerStarted","Data":"fe6322df77a7424d475750b812ebb3bacd32eea8420f037360442254cd9fae19"} Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.425585 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.522267 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts\") pod \"14458a1a-5f03-4fff-8983-1547756eb34c\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.522388 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngmfl\" (UniqueName: \"kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl\") pod \"14458a1a-5f03-4fff-8983-1547756eb34c\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.522467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data\") pod \"14458a1a-5f03-4fff-8983-1547756eb34c\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.522556 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle\") pod \"14458a1a-5f03-4fff-8983-1547756eb34c\" (UID: \"14458a1a-5f03-4fff-8983-1547756eb34c\") " Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.529265 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts" (OuterVolumeSpecName: "scripts") pod "14458a1a-5f03-4fff-8983-1547756eb34c" (UID: "14458a1a-5f03-4fff-8983-1547756eb34c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.545663 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl" (OuterVolumeSpecName: "kube-api-access-ngmfl") pod "14458a1a-5f03-4fff-8983-1547756eb34c" (UID: "14458a1a-5f03-4fff-8983-1547756eb34c"). InnerVolumeSpecName "kube-api-access-ngmfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.555840 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14458a1a-5f03-4fff-8983-1547756eb34c" (UID: "14458a1a-5f03-4fff-8983-1547756eb34c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.581292 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data" (OuterVolumeSpecName: "config-data") pod "14458a1a-5f03-4fff-8983-1547756eb34c" (UID: "14458a1a-5f03-4fff-8983-1547756eb34c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.624276 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.624498 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.624565 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14458a1a-5f03-4fff-8983-1547756eb34c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:27 crc kubenswrapper[5018]: I1014 07:10:27.624646 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngmfl\" (UniqueName: \"kubernetes.io/projected/14458a1a-5f03-4fff-8983-1547756eb34c-kube-api-access-ngmfl\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.073110 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-skccn" event={"ID":"14458a1a-5f03-4fff-8983-1547756eb34c","Type":"ContainerDied","Data":"8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6"} Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.073478 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8079777c933c04544be954d515a455ffe22ec30fe997333952a58bd25755a8c6" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.073170 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-skccn" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.212772 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:10:28 crc kubenswrapper[5018]: E1014 07:10:28.213378 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14458a1a-5f03-4fff-8983-1547756eb34c" containerName="nova-cell0-conductor-db-sync" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.213503 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="14458a1a-5f03-4fff-8983-1547756eb34c" containerName="nova-cell0-conductor-db-sync" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.214927 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="14458a1a-5f03-4fff-8983-1547756eb34c" containerName="nova-cell0-conductor-db-sync" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.215607 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.225260 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tnjd5" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.225765 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.234527 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.339066 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.339197 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.339283 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jth68\" (UniqueName: \"kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.441738 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.441932 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jth68\" (UniqueName: \"kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.442068 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.446918 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.446927 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.458218 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jth68\" (UniqueName: \"kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68\") pod \"nova-cell0-conductor-0\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:28 crc kubenswrapper[5018]: I1014 07:10:28.593872 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:29 crc kubenswrapper[5018]: I1014 07:10:29.052492 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:10:29 crc kubenswrapper[5018]: W1014 07:10:29.059254 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b076446_7046_4b7a_b315_f8a560d5604b.slice/crio-f06f3e9d0065ec29bdf951365c7634fb8922e9d418f7018bd4afa39e89b5df38 WatchSource:0}: Error finding container f06f3e9d0065ec29bdf951365c7634fb8922e9d418f7018bd4afa39e89b5df38: Status 404 returned error can't find the container with id f06f3e9d0065ec29bdf951365c7634fb8922e9d418f7018bd4afa39e89b5df38 Oct 14 07:10:29 crc kubenswrapper[5018]: I1014 07:10:29.086029 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5b076446-7046-4b7a-b315-f8a560d5604b","Type":"ContainerStarted","Data":"f06f3e9d0065ec29bdf951365c7634fb8922e9d418f7018bd4afa39e89b5df38"} Oct 14 07:10:29 crc kubenswrapper[5018]: I1014 07:10:29.091023 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerStarted","Data":"bf07b46b7cc5e2c9c4c9ce6bf95b433ad0a7c99598d56f7fc91a388368a3f53f"} Oct 14 07:10:29 crc kubenswrapper[5018]: I1014 07:10:29.130644 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.894722491 podStartE2EDuration="6.130598845s" podCreationTimestamp="2025-10-14 07:10:23 +0000 UTC" firstStartedPulling="2025-10-14 07:10:23.979956817 +0000 UTC m=+1240.564003454" lastFinishedPulling="2025-10-14 07:10:28.215833171 +0000 UTC m=+1244.799879808" observedRunningTime="2025-10-14 07:10:29.119309286 +0000 UTC m=+1245.703355943" watchObservedRunningTime="2025-10-14 07:10:29.130598845 +0000 UTC m=+1245.714645512" Oct 14 07:10:30 crc kubenswrapper[5018]: I1014 07:10:30.109758 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5b076446-7046-4b7a-b315-f8a560d5604b","Type":"ContainerStarted","Data":"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178"} Oct 14 07:10:30 crc kubenswrapper[5018]: I1014 07:10:30.110113 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:10:30 crc kubenswrapper[5018]: I1014 07:10:30.110133 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:30 crc kubenswrapper[5018]: I1014 07:10:30.133823 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.133794677 podStartE2EDuration="2.133794677s" podCreationTimestamp="2025-10-14 07:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:30.129741162 +0000 UTC m=+1246.713787799" watchObservedRunningTime="2025-10-14 07:10:30.133794677 +0000 UTC m=+1246.717841304" Oct 14 07:10:38 crc kubenswrapper[5018]: I1014 07:10:38.642448 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.230839 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lhzhc"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.231963 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.234963 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.236122 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.245883 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhzhc"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.328947 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.328993 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whc5x\" (UniqueName: \"kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.329012 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.329085 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.431231 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.431348 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.431371 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whc5x\" (UniqueName: \"kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.431386 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.437786 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.441192 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.452871 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.458540 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.460432 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.464691 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.473140 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.474233 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whc5x\" (UniqueName: \"kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x\") pod \"nova-cell0-cell-mapping-lhzhc\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.510296 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.511876 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.513888 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.532978 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7bvm\" (UniqueName: \"kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.533037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.533110 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.533155 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.533789 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.576205 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.591281 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.592777 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.597141 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.606541 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.608859 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.619411 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.620013 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.634835 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7bvm\" (UniqueName: \"kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.634956 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.635877 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.636022 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26x76\" (UniqueName: \"kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.636093 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.636156 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.636234 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.636755 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.647705 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.666323 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.674002 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.695489 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7bvm\" (UniqueName: \"kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm\") pod \"nova-api-0\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.729284 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.759847 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771352 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771403 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771444 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771482 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771515 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771540 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771603 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wzqw\" (UniqueName: \"kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771656 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26x76\" (UniqueName: \"kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771754 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.771933 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q585l\" (UniqueName: \"kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.779691 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.781279 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.795212 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26x76\" (UniqueName: \"kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76\") pod \"nova-cell1-novncproxy-0\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.824730 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873064 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q585l\" (UniqueName: \"kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873115 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873142 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873163 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873182 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873200 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873222 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873240 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873268 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873287 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wzqw\" (UniqueName: \"kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873313 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873346 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zd7t\" (UniqueName: \"kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.873367 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.874317 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.875696 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.876603 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.877931 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.882313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.885937 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.891445 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.892689 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q585l\" (UniqueName: \"kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l\") pod \"nova-scheduler-0\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.893588 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wzqw\" (UniqueName: \"kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw\") pod \"nova-metadata-0\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " pod="openstack/nova-metadata-0" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975073 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975127 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975178 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975210 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975251 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zd7t\" (UniqueName: \"kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.975274 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.976846 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.977353 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.977772 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.977888 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.978255 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:39 crc kubenswrapper[5018]: I1014 07:10:39.996188 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zd7t\" (UniqueName: \"kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t\") pod \"dnsmasq-dns-d74749bf5-9rn9c\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.041057 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.058739 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.098031 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.158398 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhzhc"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.179531 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvsxb"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.181940 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.185384 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.185589 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.191810 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvsxb"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.248697 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhzhc" event={"ID":"0f4e545d-4740-4b89-b2df-e59bf8e09144","Type":"ContainerStarted","Data":"464c34e28517b9388219c8ea515201b142b5d79099cfcc518ff84f8af70e9ef2"} Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.280765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6slvr\" (UniqueName: \"kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.280842 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.280865 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.280924 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.342631 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.388816 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6slvr\" (UniqueName: \"kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.388879 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.388898 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.388959 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.403541 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.403992 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.404075 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.406112 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6slvr\" (UniqueName: \"kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr\") pod \"nova-cell1-conductor-db-sync-tvsxb\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.486117 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.504177 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:40 crc kubenswrapper[5018]: W1014 07:10:40.618226 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44b35315_8239_4465_9a4e_6838e8fbdff8.slice/crio-363d3b4ecb3aedc8da2c7e2cb5961d6ce89faec8ed4bef78581111cdd11b95ed WatchSource:0}: Error finding container 363d3b4ecb3aedc8da2c7e2cb5961d6ce89faec8ed4bef78581111cdd11b95ed: Status 404 returned error can't find the container with id 363d3b4ecb3aedc8da2c7e2cb5961d6ce89faec8ed4bef78581111cdd11b95ed Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.630410 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.725750 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:10:40 crc kubenswrapper[5018]: I1014 07:10:40.745839 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.057094 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvsxb"] Oct 14 07:10:41 crc kubenswrapper[5018]: W1014 07:10:41.062351 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c086920_52b2_4c56_9ae6_b2c23936d783.slice/crio-2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df WatchSource:0}: Error finding container 2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df: Status 404 returned error can't find the container with id 2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.260147 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64","Type":"ContainerStarted","Data":"bfd4fd0af66099c42bbdfe04b9ece066ec9a6c77add50ddb77d99602299505ec"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.262790 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44b35315-8239-4465-9a4e-6838e8fbdff8","Type":"ContainerStarted","Data":"363d3b4ecb3aedc8da2c7e2cb5961d6ce89faec8ed4bef78581111cdd11b95ed"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.265378 5018 generic.go:334] "Generic (PLEG): container finished" podID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerID="99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83" exitCode=0 Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.265472 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" event={"ID":"f720b934-8cb2-4e78-8060-2fc0d23230fa","Type":"ContainerDied","Data":"99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.265508 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" event={"ID":"f720b934-8cb2-4e78-8060-2fc0d23230fa","Type":"ContainerStarted","Data":"bfbe18bd917131d9b9a2cca3724d880223e7e3a46b80c1c4ecdacbd2a33d68eb"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.266799 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerStarted","Data":"5ce4f9872af5f5cbb8e247809e0d38d0d90c29692098b0bb6bd178804d6d8074"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.272850 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerStarted","Data":"8979dd92661c3d00223f6753d85f8b21228de578013db2cf31d098e9cf27aefa"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.280459 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhzhc" event={"ID":"0f4e545d-4740-4b89-b2df-e59bf8e09144","Type":"ContainerStarted","Data":"27d9c4ea59e66a6c005bf07155f6c75635522244901f882a986117cbb32462c3"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.287866 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" event={"ID":"8c086920-52b2-4c56-9ae6-b2c23936d783","Type":"ContainerStarted","Data":"57ed62c426c1d9d523dff08f06246a7f633c0647bf1efed3794c0c9655b61674"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.287954 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" event={"ID":"8c086920-52b2-4c56-9ae6-b2c23936d783","Type":"ContainerStarted","Data":"2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df"} Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.316866 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lhzhc" podStartSLOduration=2.316838205 podStartE2EDuration="2.316838205s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:41.301762568 +0000 UTC m=+1257.885809195" watchObservedRunningTime="2025-10-14 07:10:41.316838205 +0000 UTC m=+1257.900884842" Oct 14 07:10:41 crc kubenswrapper[5018]: I1014 07:10:41.325311 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" podStartSLOduration=1.325294855 podStartE2EDuration="1.325294855s" podCreationTimestamp="2025-10-14 07:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:41.316712802 +0000 UTC m=+1257.900759419" watchObservedRunningTime="2025-10-14 07:10:41.325294855 +0000 UTC m=+1257.909341482" Oct 14 07:10:42 crc kubenswrapper[5018]: I1014 07:10:42.337468 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" event={"ID":"f720b934-8cb2-4e78-8060-2fc0d23230fa","Type":"ContainerStarted","Data":"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64"} Oct 14 07:10:42 crc kubenswrapper[5018]: I1014 07:10:42.338497 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:42 crc kubenswrapper[5018]: I1014 07:10:42.367140 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" podStartSLOduration=3.367122871 podStartE2EDuration="3.367122871s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:42.361871352 +0000 UTC m=+1258.945917979" watchObservedRunningTime="2025-10-14 07:10:42.367122871 +0000 UTC m=+1258.951169498" Oct 14 07:10:43 crc kubenswrapper[5018]: I1014 07:10:43.174430 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:43 crc kubenswrapper[5018]: I1014 07:10:43.211854 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.356522 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerStarted","Data":"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.356875 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerStarted","Data":"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.356983 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-log" containerID="cri-o://a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" gracePeriod=30 Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.357366 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-metadata" containerID="cri-o://e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" gracePeriod=30 Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.378000 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerStarted","Data":"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.378089 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerStarted","Data":"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.380083 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64","Type":"ContainerStarted","Data":"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.380183 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2" gracePeriod=30 Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.386588 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44b35315-8239-4465-9a4e-6838e8fbdff8","Type":"ContainerStarted","Data":"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b"} Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.399530 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.80938731 podStartE2EDuration="5.399517699s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="2025-10-14 07:10:40.758763945 +0000 UTC m=+1257.342810572" lastFinishedPulling="2025-10-14 07:10:43.348894334 +0000 UTC m=+1259.932940961" observedRunningTime="2025-10-14 07:10:44.384332168 +0000 UTC m=+1260.968378855" watchObservedRunningTime="2025-10-14 07:10:44.399517699 +0000 UTC m=+1260.983564326" Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.415059 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.430220758 podStartE2EDuration="5.415022528s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="2025-10-14 07:10:40.368926121 +0000 UTC m=+1256.952972748" lastFinishedPulling="2025-10-14 07:10:43.353727891 +0000 UTC m=+1259.937774518" observedRunningTime="2025-10-14 07:10:44.408057801 +0000 UTC m=+1260.992104478" watchObservedRunningTime="2025-10-14 07:10:44.415022528 +0000 UTC m=+1260.999069205" Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.446478 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.597184847 podStartE2EDuration="5.446459038s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="2025-10-14 07:10:40.499119079 +0000 UTC m=+1257.083165706" lastFinishedPulling="2025-10-14 07:10:43.34839327 +0000 UTC m=+1259.932439897" observedRunningTime="2025-10-14 07:10:44.439261795 +0000 UTC m=+1261.023308452" watchObservedRunningTime="2025-10-14 07:10:44.446459038 +0000 UTC m=+1261.030505665" Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.459643 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7238905879999997 podStartE2EDuration="5.459627252s" podCreationTimestamp="2025-10-14 07:10:39 +0000 UTC" firstStartedPulling="2025-10-14 07:10:40.622672499 +0000 UTC m=+1257.206719116" lastFinishedPulling="2025-10-14 07:10:43.358409153 +0000 UTC m=+1259.942455780" observedRunningTime="2025-10-14 07:10:44.458971393 +0000 UTC m=+1261.043018020" watchObservedRunningTime="2025-10-14 07:10:44.459627252 +0000 UTC m=+1261.043673879" Oct 14 07:10:44 crc kubenswrapper[5018]: I1014 07:10:44.892436 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.042481 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.060542 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.060581 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.154749 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.291783 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle\") pod \"6c354958-25bd-4422-94fb-24dd51933041\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.291892 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wzqw\" (UniqueName: \"kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw\") pod \"6c354958-25bd-4422-94fb-24dd51933041\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.291975 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs\") pod \"6c354958-25bd-4422-94fb-24dd51933041\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.292049 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data\") pod \"6c354958-25bd-4422-94fb-24dd51933041\" (UID: \"6c354958-25bd-4422-94fb-24dd51933041\") " Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.292360 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs" (OuterVolumeSpecName: "logs") pod "6c354958-25bd-4422-94fb-24dd51933041" (UID: "6c354958-25bd-4422-94fb-24dd51933041"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.292812 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c354958-25bd-4422-94fb-24dd51933041-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.302541 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw" (OuterVolumeSpecName: "kube-api-access-9wzqw") pod "6c354958-25bd-4422-94fb-24dd51933041" (UID: "6c354958-25bd-4422-94fb-24dd51933041"). InnerVolumeSpecName "kube-api-access-9wzqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.339800 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c354958-25bd-4422-94fb-24dd51933041" (UID: "6c354958-25bd-4422-94fb-24dd51933041"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.343843 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data" (OuterVolumeSpecName: "config-data") pod "6c354958-25bd-4422-94fb-24dd51933041" (UID: "6c354958-25bd-4422-94fb-24dd51933041"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.395064 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.395118 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wzqw\" (UniqueName: \"kubernetes.io/projected/6c354958-25bd-4422-94fb-24dd51933041-kube-api-access-9wzqw\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.395134 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c354958-25bd-4422-94fb-24dd51933041-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397764 5018 generic.go:334] "Generic (PLEG): container finished" podID="6c354958-25bd-4422-94fb-24dd51933041" containerID="e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" exitCode=0 Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397806 5018 generic.go:334] "Generic (PLEG): container finished" podID="6c354958-25bd-4422-94fb-24dd51933041" containerID="a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" exitCode=143 Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397869 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerDied","Data":"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c"} Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397933 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerDied","Data":"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e"} Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397947 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c354958-25bd-4422-94fb-24dd51933041","Type":"ContainerDied","Data":"5ce4f9872af5f5cbb8e247809e0d38d0d90c29692098b0bb6bd178804d6d8074"} Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397965 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.397969 5018 scope.go:117] "RemoveContainer" containerID="e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.426967 5018 scope.go:117] "RemoveContainer" containerID="a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.450475 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.465720 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.470673 5018 scope.go:117] "RemoveContainer" containerID="e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" Oct 14 07:10:45 crc kubenswrapper[5018]: E1014 07:10:45.471933 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c\": container with ID starting with e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c not found: ID does not exist" containerID="e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.471975 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c"} err="failed to get container status \"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c\": rpc error: code = NotFound desc = could not find container \"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c\": container with ID starting with e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c not found: ID does not exist" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.472028 5018 scope.go:117] "RemoveContainer" containerID="a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" Oct 14 07:10:45 crc kubenswrapper[5018]: E1014 07:10:45.472542 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e\": container with ID starting with a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e not found: ID does not exist" containerID="a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.472569 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e"} err="failed to get container status \"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e\": rpc error: code = NotFound desc = could not find container \"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e\": container with ID starting with a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e not found: ID does not exist" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.472584 5018 scope.go:117] "RemoveContainer" containerID="e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.473044 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c"} err="failed to get container status \"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c\": rpc error: code = NotFound desc = could not find container \"e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c\": container with ID starting with e112b11fb4c1084c2ccc1d636799570e2ec0bbdabf73bdddfd6885b32b6f137c not found: ID does not exist" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.473072 5018 scope.go:117] "RemoveContainer" containerID="a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.473442 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e"} err="failed to get container status \"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e\": rpc error: code = NotFound desc = could not find container \"a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e\": container with ID starting with a663d3420c096f6fa789d92498fe5baa448e87ee4d4850a3d43b5bb7ab6ad46e not found: ID does not exist" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.487279 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:45 crc kubenswrapper[5018]: E1014 07:10:45.487849 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-log" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.487900 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-log" Oct 14 07:10:45 crc kubenswrapper[5018]: E1014 07:10:45.487916 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-metadata" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.487923 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-metadata" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.488120 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-metadata" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.488171 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c354958-25bd-4422-94fb-24dd51933041" containerName="nova-metadata-log" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.489461 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.494636 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.494655 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.495978 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.599061 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.599224 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c54gx\" (UniqueName: \"kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.599267 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.599330 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.599360 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.701027 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c54gx\" (UniqueName: \"kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.701092 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.701179 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.701209 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.701248 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.702451 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.707596 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.715863 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.716163 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.723917 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c54gx\" (UniqueName: \"kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx\") pod \"nova-metadata-0\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " pod="openstack/nova-metadata-0" Oct 14 07:10:45 crc kubenswrapper[5018]: I1014 07:10:45.813162 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:46 crc kubenswrapper[5018]: I1014 07:10:46.345265 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:46 crc kubenswrapper[5018]: I1014 07:10:46.413559 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerStarted","Data":"1f86d290e9d90b8ec8263673bd00ed731dc17a9dbeff6a0ad294b363a7bbbfdf"} Oct 14 07:10:46 crc kubenswrapper[5018]: I1014 07:10:46.625488 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c354958-25bd-4422-94fb-24dd51933041" path="/var/lib/kubelet/pods/6c354958-25bd-4422-94fb-24dd51933041/volumes" Oct 14 07:10:47 crc kubenswrapper[5018]: I1014 07:10:47.433775 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f4e545d-4740-4b89-b2df-e59bf8e09144" containerID="27d9c4ea59e66a6c005bf07155f6c75635522244901f882a986117cbb32462c3" exitCode=0 Oct 14 07:10:47 crc kubenswrapper[5018]: I1014 07:10:47.434118 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhzhc" event={"ID":"0f4e545d-4740-4b89-b2df-e59bf8e09144","Type":"ContainerDied","Data":"27d9c4ea59e66a6c005bf07155f6c75635522244901f882a986117cbb32462c3"} Oct 14 07:10:47 crc kubenswrapper[5018]: I1014 07:10:47.444697 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerStarted","Data":"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172"} Oct 14 07:10:47 crc kubenswrapper[5018]: I1014 07:10:47.444786 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerStarted","Data":"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902"} Oct 14 07:10:47 crc kubenswrapper[5018]: I1014 07:10:47.528609 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.528572656 podStartE2EDuration="2.528572656s" podCreationTimestamp="2025-10-14 07:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:47.509824184 +0000 UTC m=+1264.093870821" watchObservedRunningTime="2025-10-14 07:10:47.528572656 +0000 UTC m=+1264.112619283" Oct 14 07:10:48 crc kubenswrapper[5018]: I1014 07:10:48.461124 5018 generic.go:334] "Generic (PLEG): container finished" podID="8c086920-52b2-4c56-9ae6-b2c23936d783" containerID="57ed62c426c1d9d523dff08f06246a7f633c0647bf1efed3794c0c9655b61674" exitCode=0 Oct 14 07:10:48 crc kubenswrapper[5018]: I1014 07:10:48.461439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" event={"ID":"8c086920-52b2-4c56-9ae6-b2c23936d783","Type":"ContainerDied","Data":"57ed62c426c1d9d523dff08f06246a7f633c0647bf1efed3794c0c9655b61674"} Oct 14 07:10:48 crc kubenswrapper[5018]: I1014 07:10:48.958694 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.077423 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle\") pod \"0f4e545d-4740-4b89-b2df-e59bf8e09144\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.077593 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts\") pod \"0f4e545d-4740-4b89-b2df-e59bf8e09144\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.077705 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whc5x\" (UniqueName: \"kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x\") pod \"0f4e545d-4740-4b89-b2df-e59bf8e09144\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.077752 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data\") pod \"0f4e545d-4740-4b89-b2df-e59bf8e09144\" (UID: \"0f4e545d-4740-4b89-b2df-e59bf8e09144\") " Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.085194 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x" (OuterVolumeSpecName: "kube-api-access-whc5x") pod "0f4e545d-4740-4b89-b2df-e59bf8e09144" (UID: "0f4e545d-4740-4b89-b2df-e59bf8e09144"). InnerVolumeSpecName "kube-api-access-whc5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.085678 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts" (OuterVolumeSpecName: "scripts") pod "0f4e545d-4740-4b89-b2df-e59bf8e09144" (UID: "0f4e545d-4740-4b89-b2df-e59bf8e09144"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.106293 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f4e545d-4740-4b89-b2df-e59bf8e09144" (UID: "0f4e545d-4740-4b89-b2df-e59bf8e09144"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.145098 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data" (OuterVolumeSpecName: "config-data") pod "0f4e545d-4740-4b89-b2df-e59bf8e09144" (UID: "0f4e545d-4740-4b89-b2df-e59bf8e09144"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.180044 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.180077 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.180086 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whc5x\" (UniqueName: \"kubernetes.io/projected/0f4e545d-4740-4b89-b2df-e59bf8e09144-kube-api-access-whc5x\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.180102 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4e545d-4740-4b89-b2df-e59bf8e09144-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.477155 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhzhc" event={"ID":"0f4e545d-4740-4b89-b2df-e59bf8e09144","Type":"ContainerDied","Data":"464c34e28517b9388219c8ea515201b142b5d79099cfcc518ff84f8af70e9ef2"} Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.477659 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="464c34e28517b9388219c8ea515201b142b5d79099cfcc518ff84f8af70e9ef2" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.477173 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhzhc" Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.599525 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.599891 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-api" containerID="cri-o://737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" gracePeriod=30 Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.600373 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-log" containerID="cri-o://5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" gracePeriod=30 Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.662825 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.663431 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-log" containerID="cri-o://dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" gracePeriod=30 Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.663898 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-metadata" containerID="cri-o://dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" gracePeriod=30 Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.674160 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:49 crc kubenswrapper[5018]: I1014 07:10:49.674402 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="44b35315-8239-4465-9a4e-6838e8fbdff8" containerName="nova-scheduler-scheduler" containerID="cri-o://0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b" gracePeriod=30 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.019842 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.100387 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle\") pod \"8c086920-52b2-4c56-9ae6-b2c23936d783\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.100483 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6slvr\" (UniqueName: \"kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr\") pod \"8c086920-52b2-4c56-9ae6-b2c23936d783\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.100560 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data\") pod \"8c086920-52b2-4c56-9ae6-b2c23936d783\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.100603 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts\") pod \"8c086920-52b2-4c56-9ae6-b2c23936d783\" (UID: \"8c086920-52b2-4c56-9ae6-b2c23936d783\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.100777 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.106891 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts" (OuterVolumeSpecName: "scripts") pod "8c086920-52b2-4c56-9ae6-b2c23936d783" (UID: "8c086920-52b2-4c56-9ae6-b2c23936d783"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.106931 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr" (OuterVolumeSpecName: "kube-api-access-6slvr") pod "8c086920-52b2-4c56-9ae6-b2c23936d783" (UID: "8c086920-52b2-4c56-9ae6-b2c23936d783"). InnerVolumeSpecName "kube-api-access-6slvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.174419 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c086920-52b2-4c56-9ae6-b2c23936d783" (UID: "8c086920-52b2-4c56-9ae6-b2c23936d783"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.183489 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.183803 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="dnsmasq-dns" containerID="cri-o://2d0c149e193aa09e131eec98a80bc6bb0be317bd1605f38db67dc9c18dd80398" gracePeriod=10 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.193155 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.204594 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6slvr\" (UniqueName: \"kubernetes.io/projected/8c086920-52b2-4c56-9ae6-b2c23936d783-kube-api-access-6slvr\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.204633 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.204645 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.219178 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data" (OuterVolumeSpecName: "config-data") pod "8c086920-52b2-4c56-9ae6-b2c23936d783" (UID: "8c086920-52b2-4c56-9ae6-b2c23936d783"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.265970 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.305550 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs\") pod \"81f84dc1-4504-43cc-852a-b72c6515125a\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.305649 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle\") pod \"81f84dc1-4504-43cc-852a-b72c6515125a\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.305800 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7bvm\" (UniqueName: \"kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm\") pod \"81f84dc1-4504-43cc-852a-b72c6515125a\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.305834 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data\") pod \"81f84dc1-4504-43cc-852a-b72c6515125a\" (UID: \"81f84dc1-4504-43cc-852a-b72c6515125a\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.305855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs" (OuterVolumeSpecName: "logs") pod "81f84dc1-4504-43cc-852a-b72c6515125a" (UID: "81f84dc1-4504-43cc-852a-b72c6515125a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.306408 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f84dc1-4504-43cc-852a-b72c6515125a-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.306425 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c086920-52b2-4c56-9ae6-b2c23936d783-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.308557 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm" (OuterVolumeSpecName: "kube-api-access-f7bvm") pod "81f84dc1-4504-43cc-852a-b72c6515125a" (UID: "81f84dc1-4504-43cc-852a-b72c6515125a"). InnerVolumeSpecName "kube-api-access-f7bvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.331415 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81f84dc1-4504-43cc-852a-b72c6515125a" (UID: "81f84dc1-4504-43cc-852a-b72c6515125a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.343536 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data" (OuterVolumeSpecName: "config-data") pod "81f84dc1-4504-43cc-852a-b72c6515125a" (UID: "81f84dc1-4504-43cc-852a-b72c6515125a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.407277 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data\") pod \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.407715 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c54gx\" (UniqueName: \"kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx\") pod \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.407785 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle\") pod \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.408031 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs\") pod \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.408345 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs" (OuterVolumeSpecName: "logs") pod "b651ccf7-d4d5-4310-bcaa-4aae9020dd09" (UID: "b651ccf7-d4d5-4310-bcaa-4aae9020dd09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.408521 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs\") pod \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\" (UID: \"b651ccf7-d4d5-4310-bcaa-4aae9020dd09\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.409378 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.409397 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.409410 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7bvm\" (UniqueName: \"kubernetes.io/projected/81f84dc1-4504-43cc-852a-b72c6515125a-kube-api-access-f7bvm\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.409421 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f84dc1-4504-43cc-852a-b72c6515125a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.410825 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx" (OuterVolumeSpecName: "kube-api-access-c54gx") pod "b651ccf7-d4d5-4310-bcaa-4aae9020dd09" (UID: "b651ccf7-d4d5-4310-bcaa-4aae9020dd09"). InnerVolumeSpecName "kube-api-access-c54gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.432332 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b651ccf7-d4d5-4310-bcaa-4aae9020dd09" (UID: "b651ccf7-d4d5-4310-bcaa-4aae9020dd09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.452723 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data" (OuterVolumeSpecName: "config-data") pod "b651ccf7-d4d5-4310-bcaa-4aae9020dd09" (UID: "b651ccf7-d4d5-4310-bcaa-4aae9020dd09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.477602 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b651ccf7-d4d5-4310-bcaa-4aae9020dd09" (UID: "b651ccf7-d4d5-4310-bcaa-4aae9020dd09"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495311 5018 generic.go:334] "Generic (PLEG): container finished" podID="81f84dc1-4504-43cc-852a-b72c6515125a" containerID="737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" exitCode=0 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495339 5018 generic.go:334] "Generic (PLEG): container finished" podID="81f84dc1-4504-43cc-852a-b72c6515125a" containerID="5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" exitCode=143 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495403 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerDied","Data":"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495438 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerDied","Data":"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495455 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495452 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f84dc1-4504-43cc-852a-b72c6515125a","Type":"ContainerDied","Data":"8979dd92661c3d00223f6753d85f8b21228de578013db2cf31d098e9cf27aefa"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.495495 5018 scope.go:117] "RemoveContainer" containerID="737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.499309 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.499321 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvsxb" event={"ID":"8c086920-52b2-4c56-9ae6-b2c23936d783","Type":"ContainerDied","Data":"2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.499345 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e3797b2024dcae6ceaebf305f9cbef3083129a4ffb619cb401366a244afd0df" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502730 5018 generic.go:334] "Generic (PLEG): container finished" podID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerID="dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" exitCode=0 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502765 5018 generic.go:334] "Generic (PLEG): container finished" podID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerID="dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" exitCode=143 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502813 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerDied","Data":"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502841 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerDied","Data":"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502851 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b651ccf7-d4d5-4310-bcaa-4aae9020dd09","Type":"ContainerDied","Data":"1f86d290e9d90b8ec8263673bd00ed731dc17a9dbeff6a0ad294b363a7bbbfdf"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.502903 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.511903 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.511936 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.511999 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.512008 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c54gx\" (UniqueName: \"kubernetes.io/projected/b651ccf7-d4d5-4310-bcaa-4aae9020dd09-kube-api-access-c54gx\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.515339 5018 generic.go:334] "Generic (PLEG): container finished" podID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerID="2d0c149e193aa09e131eec98a80bc6bb0be317bd1605f38db67dc9c18dd80398" exitCode=0 Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.515369 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" event={"ID":"67d09018-563a-4a79-aa5b-84ca6f7c441e","Type":"ContainerDied","Data":"2d0c149e193aa09e131eec98a80bc6bb0be317bd1605f38db67dc9c18dd80398"} Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.540281 5018 scope.go:117] "RemoveContainer" containerID="5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.586952 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.609227 5018 scope.go:117] "RemoveContainer" containerID="737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.615769 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7\": container with ID starting with 737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7 not found: ID does not exist" containerID="737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.615826 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7"} err="failed to get container status \"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7\": rpc error: code = NotFound desc = could not find container \"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7\": container with ID starting with 737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.615856 5018 scope.go:117] "RemoveContainer" containerID="5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.616225 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537\": container with ID starting with 5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537 not found: ID does not exist" containerID="5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616248 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537"} err="failed to get container status \"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537\": rpc error: code = NotFound desc = could not find container \"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537\": container with ID starting with 5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616262 5018 scope.go:117] "RemoveContainer" containerID="737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616454 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7"} err="failed to get container status \"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7\": rpc error: code = NotFound desc = could not find container \"737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7\": container with ID starting with 737798f83641a1cb69ad17eda7281785fdef34d53bdf23f0454b8733b1f970b7 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616473 5018 scope.go:117] "RemoveContainer" containerID="5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616643 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537"} err="failed to get container status \"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537\": rpc error: code = NotFound desc = could not find container \"5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537\": container with ID starting with 5bb1010f6df1be676933b95fe55a17838ce5b5dfba49f80580dcb0f318baf537 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.616656 5018 scope.go:117] "RemoveContainer" containerID="dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.632055 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.634879 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635279 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-api" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635294 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-api" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635324 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-log" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635329 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-log" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635342 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-metadata" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635347 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-metadata" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635361 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-log" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635367 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-log" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635373 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4e545d-4740-4b89-b2df-e59bf8e09144" containerName="nova-manage" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635378 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4e545d-4740-4b89-b2df-e59bf8e09144" containerName="nova-manage" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.635390 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c086920-52b2-4c56-9ae6-b2c23936d783" containerName="nova-cell1-conductor-db-sync" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635397 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c086920-52b2-4c56-9ae6-b2c23936d783" containerName="nova-cell1-conductor-db-sync" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635560 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c086920-52b2-4c56-9ae6-b2c23936d783" containerName="nova-cell1-conductor-db-sync" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635574 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4e545d-4740-4b89-b2df-e59bf8e09144" containerName="nova-manage" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635580 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-log" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635593 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" containerName="nova-api-api" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635603 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-metadata" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.635612 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" containerName="nova-metadata-log" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.636767 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.640423 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.662257 5018 scope.go:117] "RemoveContainer" containerID="dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.669430 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.684613 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.686833 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.690026 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.690196 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.692158 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.695767 5018 scope.go:117] "RemoveContainer" containerID="dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.696446 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172\": container with ID starting with dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172 not found: ID does not exist" containerID="dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.696677 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172"} err="failed to get container status \"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172\": rpc error: code = NotFound desc = could not find container \"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172\": container with ID starting with dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.696704 5018 scope.go:117] "RemoveContainer" containerID="dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.697984 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902\": container with ID starting with dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902 not found: ID does not exist" containerID="dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.698003 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902"} err="failed to get container status \"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902\": rpc error: code = NotFound desc = could not find container \"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902\": container with ID starting with dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.698018 5018 scope.go:117] "RemoveContainer" containerID="dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.700757 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.701976 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172"} err="failed to get container status \"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172\": rpc error: code = NotFound desc = could not find container \"dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172\": container with ID starting with dab9d3ee9008ac9e3b908235342582eb1e72735f0c0b533e618758dcfccc4172 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.702071 5018 scope.go:117] "RemoveContainer" containerID="dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.702394 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902"} err="failed to get container status \"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902\": rpc error: code = NotFound desc = could not find container \"dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902\": container with ID starting with dacad60e399a8e86e09826f84a202b41083561a8ca7522eac141346ee49b3902 not found: ID does not exist" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.708140 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.718658 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.719113 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="init" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.719132 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="init" Oct 14 07:10:50 crc kubenswrapper[5018]: E1014 07:10:50.719166 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="dnsmasq-dns" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.719173 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="dnsmasq-dns" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.719368 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" containerName="dnsmasq-dns" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.721811 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.722034 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxvmm\" (UniqueName: \"kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.722158 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.722192 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.722346 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.723527 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.724082 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.753349 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.823571 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.823772 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf8lj\" (UniqueName: \"kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.823932 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.824102 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.824307 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.824417 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc\") pod \"67d09018-563a-4a79-aa5b-84ca6f7c441e\" (UID: \"67d09018-563a-4a79-aa5b-84ca6f7c441e\") " Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.824957 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825232 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825383 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq9fg\" (UniqueName: \"kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825544 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825642 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825735 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxvmm\" (UniqueName: \"kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825821 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825903 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.825991 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.826139 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.826240 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.826310 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbmkj\" (UniqueName: \"kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.827749 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.828825 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj" (OuterVolumeSpecName: "kube-api-access-hf8lj") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "kube-api-access-hf8lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.836494 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.843324 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.852661 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxvmm\" (UniqueName: \"kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm\") pod \"nova-api-0\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " pod="openstack/nova-api-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.876556 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.877708 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.879423 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config" (OuterVolumeSpecName: "config") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.893936 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.906918 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67d09018-563a-4a79-aa5b-84ca6f7c441e" (UID: "67d09018-563a-4a79-aa5b-84ca6f7c441e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927380 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq9fg\" (UniqueName: \"kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927439 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927457 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927478 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927516 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927538 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.927552 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbmkj\" (UniqueName: \"kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928009 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928076 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928086 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928097 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928105 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928114 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf8lj\" (UniqueName: \"kubernetes.io/projected/67d09018-563a-4a79-aa5b-84ca6f7c441e-kube-api-access-hf8lj\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928123 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d09018-563a-4a79-aa5b-84ca6f7c441e-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.928449 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.943471 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.943791 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.943901 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.945469 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.946677 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.949072 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq9fg\" (UniqueName: \"kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg\") pod \"nova-metadata-0\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " pod="openstack/nova-metadata-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.949133 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbmkj\" (UniqueName: \"kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj\") pod \"nova-cell1-conductor-0\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:50 crc kubenswrapper[5018]: I1014 07:10:50.964452 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.005719 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.107228 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.116392 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.240359 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle\") pod \"44b35315-8239-4465-9a4e-6838e8fbdff8\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.240416 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data\") pod \"44b35315-8239-4465-9a4e-6838e8fbdff8\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.240551 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q585l\" (UniqueName: \"kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l\") pod \"44b35315-8239-4465-9a4e-6838e8fbdff8\" (UID: \"44b35315-8239-4465-9a4e-6838e8fbdff8\") " Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.254862 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l" (OuterVolumeSpecName: "kube-api-access-q585l") pod "44b35315-8239-4465-9a4e-6838e8fbdff8" (UID: "44b35315-8239-4465-9a4e-6838e8fbdff8"). InnerVolumeSpecName "kube-api-access-q585l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.272940 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44b35315-8239-4465-9a4e-6838e8fbdff8" (UID: "44b35315-8239-4465-9a4e-6838e8fbdff8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.276070 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data" (OuterVolumeSpecName: "config-data") pod "44b35315-8239-4465-9a4e-6838e8fbdff8" (UID: "44b35315-8239-4465-9a4e-6838e8fbdff8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.342307 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q585l\" (UniqueName: \"kubernetes.io/projected/44b35315-8239-4465-9a4e-6838e8fbdff8-kube-api-access-q585l\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.342336 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.342349 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b35315-8239-4465-9a4e-6838e8fbdff8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.435954 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.545446 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.546034 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb9f44c77-w69fd" event={"ID":"67d09018-563a-4a79-aa5b-84ca6f7c441e","Type":"ContainerDied","Data":"1d742d1b1a011113357dbe71e47d89043ec4cd1369d00dbf2923cac1735ca4f8"} Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.546938 5018 scope.go:117] "RemoveContainer" containerID="2d0c149e193aa09e131eec98a80bc6bb0be317bd1605f38db67dc9c18dd80398" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.549546 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerStarted","Data":"2c0bffcb10ed087a75fdb7965b076e0b7b2b5a0c3f9b5e51b5df6c9f40052c91"} Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.551815 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.554715 5018 generic.go:334] "Generic (PLEG): container finished" podID="44b35315-8239-4465-9a4e-6838e8fbdff8" containerID="0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b" exitCode=0 Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.554746 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44b35315-8239-4465-9a4e-6838e8fbdff8","Type":"ContainerDied","Data":"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b"} Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.554765 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44b35315-8239-4465-9a4e-6838e8fbdff8","Type":"ContainerDied","Data":"363d3b4ecb3aedc8da2c7e2cb5961d6ce89faec8ed4bef78581111cdd11b95ed"} Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.554814 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: W1014 07:10:51.559852 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63753948_7b37_4eeb_a378_2947a6caa55c.slice/crio-f1ae4098736e7c2a3598fd209ec60614ae6c1bb8f7876d466a4f1858d8ddd2ae WatchSource:0}: Error finding container f1ae4098736e7c2a3598fd209ec60614ae6c1bb8f7876d466a4f1858d8ddd2ae: Status 404 returned error can't find the container with id f1ae4098736e7c2a3598fd209ec60614ae6c1bb8f7876d466a4f1858d8ddd2ae Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.580102 5018 scope.go:117] "RemoveContainer" containerID="5d3f7f51b4162c7d00450085a977315fd391d7960354fba47bf2850f1ad13c4a" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.593935 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.605631 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb9f44c77-w69fd"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.626386 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.629873 5018 scope.go:117] "RemoveContainer" containerID="0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.641316 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.650171 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.657630 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: E1014 07:10:51.658003 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b35315-8239-4465-9a4e-6838e8fbdff8" containerName="nova-scheduler-scheduler" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.658017 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b35315-8239-4465-9a4e-6838e8fbdff8" containerName="nova-scheduler-scheduler" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.658179 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b35315-8239-4465-9a4e-6838e8fbdff8" containerName="nova-scheduler-scheduler" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.658965 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.661431 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.665134 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.676259 5018 scope.go:117] "RemoveContainer" containerID="0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b" Oct 14 07:10:51 crc kubenswrapper[5018]: E1014 07:10:51.676845 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b\": container with ID starting with 0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b not found: ID does not exist" containerID="0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.676875 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b"} err="failed to get container status \"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b\": rpc error: code = NotFound desc = could not find container \"0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b\": container with ID starting with 0732da4f78e618f50bea97a87fba817f6f20e1f4aedd8f2485dce57e78f36f2b not found: ID does not exist" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.750734 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.751164 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.751298 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg9d7\" (UniqueName: \"kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.856345 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.856528 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.856564 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg9d7\" (UniqueName: \"kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.860406 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.863077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.876224 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg9d7\" (UniqueName: \"kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7\") pod \"nova-scheduler-0\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " pod="openstack/nova-scheduler-0" Oct 14 07:10:51 crc kubenswrapper[5018]: I1014 07:10:51.982907 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:10:52 crc kubenswrapper[5018]: W1014 07:10:52.452037 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda42b8d6a_9b7f_4457_b8db_32cf7d1b8bf8.slice/crio-85229297bddc723820b9e066fc49cee113ad2efcc452d13cdce8e927b8c6d896 WatchSource:0}: Error finding container 85229297bddc723820b9e066fc49cee113ad2efcc452d13cdce8e927b8c6d896: Status 404 returned error can't find the container with id 85229297bddc723820b9e066fc49cee113ad2efcc452d13cdce8e927b8c6d896 Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.456064 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.592511 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerStarted","Data":"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.592566 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerStarted","Data":"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.595005 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8","Type":"ContainerStarted","Data":"85229297bddc723820b9e066fc49cee113ad2efcc452d13cdce8e927b8c6d896"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.598109 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerStarted","Data":"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.598137 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerStarted","Data":"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.598150 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerStarted","Data":"93a501a273d9d68cd6145592e5c387fb43334ddb7fe7deb007bf4d7bc950e7f1"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.600120 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"63753948-7b37-4eeb-a378-2947a6caa55c","Type":"ContainerStarted","Data":"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.600167 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"63753948-7b37-4eeb-a378-2947a6caa55c","Type":"ContainerStarted","Data":"f1ae4098736e7c2a3598fd209ec60614ae6c1bb8f7876d466a4f1858d8ddd2ae"} Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.600336 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.620492 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b35315-8239-4465-9a4e-6838e8fbdff8" path="/var/lib/kubelet/pods/44b35315-8239-4465-9a4e-6838e8fbdff8/volumes" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.621518 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d09018-563a-4a79-aa5b-84ca6f7c441e" path="/var/lib/kubelet/pods/67d09018-563a-4a79-aa5b-84ca6f7c441e/volumes" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.622300 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f84dc1-4504-43cc-852a-b72c6515125a" path="/var/lib/kubelet/pods/81f84dc1-4504-43cc-852a-b72c6515125a/volumes" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.623340 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b651ccf7-d4d5-4310-bcaa-4aae9020dd09" path="/var/lib/kubelet/pods/b651ccf7-d4d5-4310-bcaa-4aae9020dd09/volumes" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.628092 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.628061626 podStartE2EDuration="2.628061626s" podCreationTimestamp="2025-10-14 07:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:52.618562257 +0000 UTC m=+1269.202608894" watchObservedRunningTime="2025-10-14 07:10:52.628061626 +0000 UTC m=+1269.212108293" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.643566 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.643547335 podStartE2EDuration="2.643547335s" podCreationTimestamp="2025-10-14 07:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:52.638402189 +0000 UTC m=+1269.222448826" watchObservedRunningTime="2025-10-14 07:10:52.643547335 +0000 UTC m=+1269.227593962" Oct 14 07:10:52 crc kubenswrapper[5018]: I1014 07:10:52.657586 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.657570302 podStartE2EDuration="2.657570302s" podCreationTimestamp="2025-10-14 07:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:52.655956066 +0000 UTC m=+1269.240002693" watchObservedRunningTime="2025-10-14 07:10:52.657570302 +0000 UTC m=+1269.241616929" Oct 14 07:10:53 crc kubenswrapper[5018]: I1014 07:10:53.459163 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 07:10:53 crc kubenswrapper[5018]: I1014 07:10:53.622092 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8","Type":"ContainerStarted","Data":"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae"} Oct 14 07:10:53 crc kubenswrapper[5018]: I1014 07:10:53.649175 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.649153644 podStartE2EDuration="2.649153644s" podCreationTimestamp="2025-10-14 07:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:10:53.638992816 +0000 UTC m=+1270.223039493" watchObservedRunningTime="2025-10-14 07:10:53.649153644 +0000 UTC m=+1270.233200281" Oct 14 07:10:56 crc kubenswrapper[5018]: I1014 07:10:56.061933 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 07:10:56 crc kubenswrapper[5018]: I1014 07:10:56.117318 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:10:56 crc kubenswrapper[5018]: I1014 07:10:56.117601 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:10:56 crc kubenswrapper[5018]: I1014 07:10:56.983558 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:10:57 crc kubenswrapper[5018]: I1014 07:10:57.637881 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:57 crc kubenswrapper[5018]: I1014 07:10:57.638455 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" containerName="kube-state-metrics" containerID="cri-o://00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1" gracePeriod=30 Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.147632 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.287143 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqpfg\" (UniqueName: \"kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg\") pod \"25658566-e65b-48fe-aa02-1df2ab82b526\" (UID: \"25658566-e65b-48fe-aa02-1df2ab82b526\") " Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.294563 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg" (OuterVolumeSpecName: "kube-api-access-zqpfg") pod "25658566-e65b-48fe-aa02-1df2ab82b526" (UID: "25658566-e65b-48fe-aa02-1df2ab82b526"). InnerVolumeSpecName "kube-api-access-zqpfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.390140 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqpfg\" (UniqueName: \"kubernetes.io/projected/25658566-e65b-48fe-aa02-1df2ab82b526-kube-api-access-zqpfg\") on node \"crc\" DevicePath \"\"" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.670230 5018 generic.go:334] "Generic (PLEG): container finished" podID="25658566-e65b-48fe-aa02-1df2ab82b526" containerID="00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1" exitCode=2 Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.670278 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25658566-e65b-48fe-aa02-1df2ab82b526","Type":"ContainerDied","Data":"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1"} Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.670314 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25658566-e65b-48fe-aa02-1df2ab82b526","Type":"ContainerDied","Data":"9716a1206a89523fa2008c6f7ef3acd6907d2afe88e152b28ff8ef2c8a27e5aa"} Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.670311 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.670332 5018 scope.go:117] "RemoveContainer" containerID="00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.705835 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.706744 5018 scope.go:117] "RemoveContainer" containerID="00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1" Oct 14 07:10:58 crc kubenswrapper[5018]: E1014 07:10:58.707156 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1\": container with ID starting with 00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1 not found: ID does not exist" containerID="00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.707204 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1"} err="failed to get container status \"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1\": rpc error: code = NotFound desc = could not find container \"00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1\": container with ID starting with 00113753a92c4d4838bdf9466a9c1905076508c843889330bf6003588f7ed5e1 not found: ID does not exist" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.726128 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.744920 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:58 crc kubenswrapper[5018]: E1014 07:10:58.745599 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" containerName="kube-state-metrics" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.745664 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" containerName="kube-state-metrics" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.746024 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" containerName="kube-state-metrics" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.747402 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.749564 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.750008 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.758582 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.798680 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.798998 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.799162 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rqwl\" (UniqueName: \"kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.799255 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.900911 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rqwl\" (UniqueName: \"kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.900999 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.901204 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.901365 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.906749 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.906782 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.913676 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:58 crc kubenswrapper[5018]: I1014 07:10:58.924968 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rqwl\" (UniqueName: \"kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl\") pod \"kube-state-metrics-0\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " pod="openstack/kube-state-metrics-0" Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.069569 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.392327 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.393298 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-central-agent" containerID="cri-o://78d244c57eec48f01d053b75b6b15df44756f1a86d158f3abd8d35b740e667d8" gracePeriod=30 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.393325 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="proxy-httpd" containerID="cri-o://bf07b46b7cc5e2c9c4c9ce6bf95b433ad0a7c99598d56f7fc91a388368a3f53f" gracePeriod=30 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.393414 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="sg-core" containerID="cri-o://fe6322df77a7424d475750b812ebb3bacd32eea8420f037360442254cd9fae19" gracePeriod=30 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.393441 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-notification-agent" containerID="cri-o://0b41efaf38effb59e9892333dafa31b2432c64f40e8aa04ffd1d117c8bf57289" gracePeriod=30 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.562277 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:10:59 crc kubenswrapper[5018]: W1014 07:10:59.573356 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66a64e8d_ccfb_4f2c_8ee8_b509d7ec4bbe.slice/crio-fdc65b82ae3f4227b60de78688b45336f41c5f90388edef59fe4b3514d8cc591 WatchSource:0}: Error finding container fdc65b82ae3f4227b60de78688b45336f41c5f90388edef59fe4b3514d8cc591: Status 404 returned error can't find the container with id fdc65b82ae3f4227b60de78688b45336f41c5f90388edef59fe4b3514d8cc591 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.699180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe","Type":"ContainerStarted","Data":"fdc65b82ae3f4227b60de78688b45336f41c5f90388edef59fe4b3514d8cc591"} Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.712169 5018 generic.go:334] "Generic (PLEG): container finished" podID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerID="bf07b46b7cc5e2c9c4c9ce6bf95b433ad0a7c99598d56f7fc91a388368a3f53f" exitCode=0 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.712224 5018 generic.go:334] "Generic (PLEG): container finished" podID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerID="fe6322df77a7424d475750b812ebb3bacd32eea8420f037360442254cd9fae19" exitCode=2 Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.712248 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerDied","Data":"bf07b46b7cc5e2c9c4c9ce6bf95b433ad0a7c99598d56f7fc91a388368a3f53f"} Oct 14 07:10:59 crc kubenswrapper[5018]: I1014 07:10:59.712301 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerDied","Data":"fe6322df77a7424d475750b812ebb3bacd32eea8420f037360442254cd9fae19"} Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.618468 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25658566-e65b-48fe-aa02-1df2ab82b526" path="/var/lib/kubelet/pods/25658566-e65b-48fe-aa02-1df2ab82b526/volumes" Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.731210 5018 generic.go:334] "Generic (PLEG): container finished" podID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerID="78d244c57eec48f01d053b75b6b15df44756f1a86d158f3abd8d35b740e667d8" exitCode=0 Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.731288 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerDied","Data":"78d244c57eec48f01d053b75b6b15df44756f1a86d158f3abd8d35b740e667d8"} Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.733911 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe","Type":"ContainerStarted","Data":"fc84e4f07631283548a442b12f788b0fb870e48754176bcb037520224ef0cee4"} Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.734027 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.770425 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.38983648 podStartE2EDuration="2.770395651s" podCreationTimestamp="2025-10-14 07:10:58 +0000 UTC" firstStartedPulling="2025-10-14 07:10:59.575103618 +0000 UTC m=+1276.159150265" lastFinishedPulling="2025-10-14 07:10:59.955662799 +0000 UTC m=+1276.539709436" observedRunningTime="2025-10-14 07:11:00.754247694 +0000 UTC m=+1277.338294361" watchObservedRunningTime="2025-10-14 07:11:00.770395651 +0000 UTC m=+1277.354442318" Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.965357 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:00 crc kubenswrapper[5018]: I1014 07:11:00.965663 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.117385 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.117462 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.745760 5018 generic.go:334] "Generic (PLEG): container finished" podID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerID="0b41efaf38effb59e9892333dafa31b2432c64f40e8aa04ffd1d117c8bf57289" exitCode=0 Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.746878 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerDied","Data":"0b41efaf38effb59e9892333dafa31b2432c64f40e8aa04ffd1d117c8bf57289"} Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.746903 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37976e65-9d86-4384-adcc-8e4f0cee52c7","Type":"ContainerDied","Data":"e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27"} Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.746914 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c736f4248ee4c56d1afc7865181074550d7630120e368dd9d313abb2ea6d27" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.789222 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870186 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870281 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870341 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870365 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870396 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870447 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.870522 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gntcn\" (UniqueName: \"kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn\") pod \"37976e65-9d86-4384-adcc-8e4f0cee52c7\" (UID: \"37976e65-9d86-4384-adcc-8e4f0cee52c7\") " Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.871000 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.871153 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.876434 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn" (OuterVolumeSpecName: "kube-api-access-gntcn") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "kube-api-access-gntcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.877610 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts" (OuterVolumeSpecName: "scripts") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.902829 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.945773 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.965501 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972386 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972418 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37976e65-9d86-4384-adcc-8e4f0cee52c7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972429 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gntcn\" (UniqueName: \"kubernetes.io/projected/37976e65-9d86-4384-adcc-8e4f0cee52c7-kube-api-access-gntcn\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972440 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972481 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.972491 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.974038 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data" (OuterVolumeSpecName: "config-data") pod "37976e65-9d86-4384-adcc-8e4f0cee52c7" (UID: "37976e65-9d86-4384-adcc-8e4f0cee52c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:01 crc kubenswrapper[5018]: I1014 07:11:01.983394 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.007921 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.009701 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.074994 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37976e65-9d86-4384-adcc-8e4f0cee52c7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.130813 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.130762 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.754353 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.780749 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.791016 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.796701 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.815448 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:02 crc kubenswrapper[5018]: E1014 07:11:02.815885 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="sg-core" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.815902 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="sg-core" Oct 14 07:11:02 crc kubenswrapper[5018]: E1014 07:11:02.815921 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="proxy-httpd" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.815928 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="proxy-httpd" Oct 14 07:11:02 crc kubenswrapper[5018]: E1014 07:11:02.815940 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-notification-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.815948 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-notification-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: E1014 07:11:02.815958 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-central-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.815964 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-central-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.816125 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-central-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.816138 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="proxy-httpd" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.816155 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="ceilometer-notification-agent" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.816164 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" containerName="sg-core" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.817740 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.824602 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.824905 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.833081 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.850685 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.889020 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.889380 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xffgq\" (UniqueName: \"kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.889548 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.889936 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.890094 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.890214 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.890371 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.890476 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.992590 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xffgq\" (UniqueName: \"kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.993468 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.993684 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.993887 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.994072 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.994142 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.994416 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.994290 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.994770 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.995038 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.999092 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.999482 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:02 crc kubenswrapper[5018]: I1014 07:11:02.999817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.000482 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.022997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.025077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xffgq\" (UniqueName: \"kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq\") pod \"ceilometer-0\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " pod="openstack/ceilometer-0" Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.150454 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.619104 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:03 crc kubenswrapper[5018]: W1014 07:11:03.631562 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aa6aaea_328b_4b70_ada7_e88dfcd098c3.slice/crio-97912b054cf92f6a0762227970197c3c2e955b12769b0a6393812e52729e89b2 WatchSource:0}: Error finding container 97912b054cf92f6a0762227970197c3c2e955b12769b0a6393812e52729e89b2: Status 404 returned error can't find the container with id 97912b054cf92f6a0762227970197c3c2e955b12769b0a6393812e52729e89b2 Oct 14 07:11:03 crc kubenswrapper[5018]: I1014 07:11:03.767172 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerStarted","Data":"97912b054cf92f6a0762227970197c3c2e955b12769b0a6393812e52729e89b2"} Oct 14 07:11:04 crc kubenswrapper[5018]: I1014 07:11:04.623091 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37976e65-9d86-4384-adcc-8e4f0cee52c7" path="/var/lib/kubelet/pods/37976e65-9d86-4384-adcc-8e4f0cee52c7/volumes" Oct 14 07:11:04 crc kubenswrapper[5018]: I1014 07:11:04.778321 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerStarted","Data":"09218b01963ed88b7189398f1ff7bb40bcc7c3921fe4cdf3fc24bb6e3cc63da1"} Oct 14 07:11:05 crc kubenswrapper[5018]: I1014 07:11:05.792884 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerStarted","Data":"a71e05311d9e9d487e6b68af73733c475d8abdb46e34cca68a5bf7425ac670ea"} Oct 14 07:11:06 crc kubenswrapper[5018]: I1014 07:11:06.831706 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerStarted","Data":"255c2f2501dfcd526b479e9502645703d3a7e40d65978238ee9333a4df98dc99"} Oct 14 07:11:08 crc kubenswrapper[5018]: I1014 07:11:08.863675 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerStarted","Data":"97fcbd659db9eb866c49c1e9a439e1d4788e9d38fa87f88d2d7b31a2ebeead1e"} Oct 14 07:11:08 crc kubenswrapper[5018]: I1014 07:11:08.864129 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:11:08 crc kubenswrapper[5018]: I1014 07:11:08.902939 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.553229679 podStartE2EDuration="6.902916417s" podCreationTimestamp="2025-10-14 07:11:02 +0000 UTC" firstStartedPulling="2025-10-14 07:11:03.634464271 +0000 UTC m=+1280.218510898" lastFinishedPulling="2025-10-14 07:11:07.984150969 +0000 UTC m=+1284.568197636" observedRunningTime="2025-10-14 07:11:08.897465342 +0000 UTC m=+1285.481512029" watchObservedRunningTime="2025-10-14 07:11:08.902916417 +0000 UTC m=+1285.486963084" Oct 14 07:11:09 crc kubenswrapper[5018]: I1014 07:11:09.089946 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 07:11:10 crc kubenswrapper[5018]: I1014 07:11:10.973805 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:11:10 crc kubenswrapper[5018]: I1014 07:11:10.974906 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:11:10 crc kubenswrapper[5018]: I1014 07:11:10.979591 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:11:10 crc kubenswrapper[5018]: I1014 07:11:10.985017 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.127194 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.128110 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.143670 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.892571 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.895384 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:11:11 crc kubenswrapper[5018]: I1014 07:11:11.897013 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.060636 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.064199 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.097781 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130612 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130682 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130709 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130779 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130805 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gptwh\" (UniqueName: \"kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.130898 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.232041 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.232813 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.232896 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gptwh\" (UniqueName: \"kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.233295 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.233852 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.233970 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.234004 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.234737 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.235214 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.235351 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.236062 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.263280 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gptwh\" (UniqueName: \"kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh\") pod \"dnsmasq-dns-65bf758599-jh97x\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.394433 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:12 crc kubenswrapper[5018]: I1014 07:11:12.921368 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:11:12 crc kubenswrapper[5018]: W1014 07:11:12.926110 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc288e96_07a9_43b9_b153_a9246dd704eb.slice/crio-1a9a8169647985e3e2103f2369d85ad8dffb90f5d9b28247131f7be1740fd5f2 WatchSource:0}: Error finding container 1a9a8169647985e3e2103f2369d85ad8dffb90f5d9b28247131f7be1740fd5f2: Status 404 returned error can't find the container with id 1a9a8169647985e3e2103f2369d85ad8dffb90f5d9b28247131f7be1740fd5f2 Oct 14 07:11:13 crc kubenswrapper[5018]: I1014 07:11:13.910587 5018 generic.go:334] "Generic (PLEG): container finished" podID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerID="e0a7b1e3d67687b82209dd21e8cb36d14ba20c1a7b9089bb0350eb8b455c87e7" exitCode=0 Oct 14 07:11:13 crc kubenswrapper[5018]: I1014 07:11:13.913065 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf758599-jh97x" event={"ID":"dc288e96-07a9-43b9-b153-a9246dd704eb","Type":"ContainerDied","Data":"e0a7b1e3d67687b82209dd21e8cb36d14ba20c1a7b9089bb0350eb8b455c87e7"} Oct 14 07:11:13 crc kubenswrapper[5018]: I1014 07:11:13.913116 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf758599-jh97x" event={"ID":"dc288e96-07a9-43b9-b153-a9246dd704eb","Type":"ContainerStarted","Data":"1a9a8169647985e3e2103f2369d85ad8dffb90f5d9b28247131f7be1740fd5f2"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.646114 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.646870 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-central-agent" containerID="cri-o://09218b01963ed88b7189398f1ff7bb40bcc7c3921fe4cdf3fc24bb6e3cc63da1" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.647295 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="proxy-httpd" containerID="cri-o://97fcbd659db9eb866c49c1e9a439e1d4788e9d38fa87f88d2d7b31a2ebeead1e" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.647356 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="sg-core" containerID="cri-o://255c2f2501dfcd526b479e9502645703d3a7e40d65978238ee9333a4df98dc99" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.647398 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-notification-agent" containerID="cri-o://a71e05311d9e9d487e6b68af73733c475d8abdb46e34cca68a5bf7425ac670ea" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.812023 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.886315 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle\") pod \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.886533 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data\") pod \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.886723 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26x76\" (UniqueName: \"kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76\") pod \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\" (UID: \"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64\") " Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.897831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76" (OuterVolumeSpecName: "kube-api-access-26x76") pod "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" (UID: "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64"). InnerVolumeSpecName "kube-api-access-26x76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.919544 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.934338 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data" (OuterVolumeSpecName: "config-data") pod "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" (UID: "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.937691 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf758599-jh97x" event={"ID":"dc288e96-07a9-43b9-b153-a9246dd704eb","Type":"ContainerStarted","Data":"f563d6fe835cc693edf98b5acc197853d2d8d196613cf1c29f3ee82afbf51623"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.938398 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.941053 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" containerID="e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2" exitCode=137 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.941114 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64","Type":"ContainerDied","Data":"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.941140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64","Type":"ContainerDied","Data":"bfd4fd0af66099c42bbdfe04b9ece066ec9a6c77add50ddb77d99602299505ec"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.941161 5018 scope.go:117] "RemoveContainer" containerID="e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.941413 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.954325 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" (UID: "6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.960248 5018 generic.go:334] "Generic (PLEG): container finished" podID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerID="97fcbd659db9eb866c49c1e9a439e1d4788e9d38fa87f88d2d7b31a2ebeead1e" exitCode=0 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.960285 5018 generic.go:334] "Generic (PLEG): container finished" podID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerID="255c2f2501dfcd526b479e9502645703d3a7e40d65978238ee9333a4df98dc99" exitCode=2 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.960384 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerDied","Data":"97fcbd659db9eb866c49c1e9a439e1d4788e9d38fa87f88d2d7b31a2ebeead1e"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.960473 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerDied","Data":"255c2f2501dfcd526b479e9502645703d3a7e40d65978238ee9333a4df98dc99"} Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.961352 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-log" containerID="cri-o://44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.961532 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65bf758599-jh97x" podStartSLOduration=2.961506999 podStartE2EDuration="2.961506999s" podCreationTimestamp="2025-10-14 07:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:14.960083459 +0000 UTC m=+1291.544130106" watchObservedRunningTime="2025-10-14 07:11:14.961506999 +0000 UTC m=+1291.545553616" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.961783 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-api" containerID="cri-o://6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba" gracePeriod=30 Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.989326 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.989359 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26x76\" (UniqueName: \"kubernetes.io/projected/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-kube-api-access-26x76\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:14 crc kubenswrapper[5018]: I1014 07:11:14.989373 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.013917 5018 scope.go:117] "RemoveContainer" containerID="e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2" Oct 14 07:11:15 crc kubenswrapper[5018]: E1014 07:11:15.014466 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2\": container with ID starting with e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2 not found: ID does not exist" containerID="e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.014509 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2"} err="failed to get container status \"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2\": rpc error: code = NotFound desc = could not find container \"e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2\": container with ID starting with e6f4d6eecb5bc22aa059bf3c7c128bc7de2461ef9f6f124abefa792daa3e80c2 not found: ID does not exist" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.290491 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.303634 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.318246 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:11:15 crc kubenswrapper[5018]: E1014 07:11:15.318722 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.318743 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.319009 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.319943 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.321901 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.322596 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.322658 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.346279 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.395995 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.396045 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g597\" (UniqueName: \"kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.396265 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.396388 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.396433 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.498448 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g597\" (UniqueName: \"kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.498579 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.498666 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.498706 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.498762 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.502807 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.504412 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.505052 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.505523 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.517091 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g597\" (UniqueName: \"kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597\") pod \"nova-cell1-novncproxy-0\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.646459 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.974981 5018 generic.go:334] "Generic (PLEG): container finished" podID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerID="09218b01963ed88b7189398f1ff7bb40bcc7c3921fe4cdf3fc24bb6e3cc63da1" exitCode=0 Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.975085 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerDied","Data":"09218b01963ed88b7189398f1ff7bb40bcc7c3921fe4cdf3fc24bb6e3cc63da1"} Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.977147 5018 generic.go:334] "Generic (PLEG): container finished" podID="d63db659-8460-4259-9396-80cd4489c0e3" containerID="44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72" exitCode=143 Oct 14 07:11:15 crc kubenswrapper[5018]: I1014 07:11:15.977231 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerDied","Data":"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72"} Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.105316 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:11:16 crc kubenswrapper[5018]: W1014 07:11:16.110489 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53392f65_7a1b_49fc_96a4_341f237e9288.slice/crio-1cd1ec3a2c7a0f89469209f8b87ebe139e239dacdf3ec18968d24c7c7bb584f7 WatchSource:0}: Error finding container 1cd1ec3a2c7a0f89469209f8b87ebe139e239dacdf3ec18968d24c7c7bb584f7: Status 404 returned error can't find the container with id 1cd1ec3a2c7a0f89469209f8b87ebe139e239dacdf3ec18968d24c7c7bb584f7 Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.629789 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64" path="/var/lib/kubelet/pods/6ad4eb41-85f5-4a5a-a22f-fe3a8a184e64/volumes" Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.987119 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"53392f65-7a1b-49fc-96a4-341f237e9288","Type":"ContainerStarted","Data":"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447"} Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.987175 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"53392f65-7a1b-49fc-96a4-341f237e9288","Type":"ContainerStarted","Data":"1cd1ec3a2c7a0f89469209f8b87ebe139e239dacdf3ec18968d24c7c7bb584f7"} Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.992072 5018 generic.go:334] "Generic (PLEG): container finished" podID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerID="a71e05311d9e9d487e6b68af73733c475d8abdb46e34cca68a5bf7425ac670ea" exitCode=0 Oct 14 07:11:16 crc kubenswrapper[5018]: I1014 07:11:16.992371 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerDied","Data":"a71e05311d9e9d487e6b68af73733c475d8abdb46e34cca68a5bf7425ac670ea"} Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.312847 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.336481 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.336452932 podStartE2EDuration="2.336452932s" podCreationTimestamp="2025-10-14 07:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:17.014673436 +0000 UTC m=+1293.598720073" watchObservedRunningTime="2025-10-14 07:11:17.336452932 +0000 UTC m=+1293.920499559" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.449680 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450711 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450809 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450862 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450914 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450944 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.450990 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xffgq\" (UniqueName: \"kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.451031 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml\") pod \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\" (UID: \"3aa6aaea-328b-4b70-ada7-e88dfcd098c3\") " Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.452920 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.459347 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.464901 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts" (OuterVolumeSpecName: "scripts") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.478357 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq" (OuterVolumeSpecName: "kube-api-access-xffgq") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "kube-api-access-xffgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.501368 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.505056 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.533908 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552928 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552957 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552966 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552975 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xffgq\" (UniqueName: \"kubernetes.io/projected/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-kube-api-access-xffgq\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552985 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.552994 5018 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.553004 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.622626 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data" (OuterVolumeSpecName: "config-data") pod "3aa6aaea-328b-4b70-ada7-e88dfcd098c3" (UID: "3aa6aaea-328b-4b70-ada7-e88dfcd098c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:17 crc kubenswrapper[5018]: I1014 07:11:17.655193 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa6aaea-328b-4b70-ada7-e88dfcd098c3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.008025 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.008843 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3aa6aaea-328b-4b70-ada7-e88dfcd098c3","Type":"ContainerDied","Data":"97912b054cf92f6a0762227970197c3c2e955b12769b0a6393812e52729e89b2"} Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.008913 5018 scope.go:117] "RemoveContainer" containerID="97fcbd659db9eb866c49c1e9a439e1d4788e9d38fa87f88d2d7b31a2ebeead1e" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.062986 5018 scope.go:117] "RemoveContainer" containerID="255c2f2501dfcd526b479e9502645703d3a7e40d65978238ee9333a4df98dc99" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.074132 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.095010 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.104178 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:18 crc kubenswrapper[5018]: E1014 07:11:18.104742 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="proxy-httpd" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.104758 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="proxy-httpd" Oct 14 07:11:18 crc kubenswrapper[5018]: E1014 07:11:18.104789 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-central-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.104797 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-central-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: E1014 07:11:18.104818 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="sg-core" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.104827 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="sg-core" Oct 14 07:11:18 crc kubenswrapper[5018]: E1014 07:11:18.104844 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-notification-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.104852 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-notification-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.105079 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="proxy-httpd" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.105108 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-central-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.105124 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="ceilometer-notification-agent" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.105136 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" containerName="sg-core" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.107317 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.111039 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.111068 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.111867 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.117324 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.181852 5018 scope.go:117] "RemoveContainer" containerID="a71e05311d9e9d487e6b68af73733c475d8abdb46e34cca68a5bf7425ac670ea" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.205878 5018 scope.go:117] "RemoveContainer" containerID="09218b01963ed88b7189398f1ff7bb40bcc7c3921fe4cdf3fc24bb6e3cc63da1" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270139 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270215 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gjp2\" (UniqueName: \"kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270247 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270311 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270367 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270390 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270423 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.270458 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.371938 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.372422 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.372443 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.372475 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.372500 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.372556 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.373098 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.373152 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gjp2\" (UniqueName: \"kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.373325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.375807 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.379409 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.379522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.379890 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.386849 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.388178 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.390697 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gjp2\" (UniqueName: \"kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.391405 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts\") pod \"ceilometer-0\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.477009 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle\") pod \"d63db659-8460-4259-9396-80cd4489c0e3\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.477092 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data\") pod \"d63db659-8460-4259-9396-80cd4489c0e3\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.477294 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxvmm\" (UniqueName: \"kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm\") pod \"d63db659-8460-4259-9396-80cd4489c0e3\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.477347 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs\") pod \"d63db659-8460-4259-9396-80cd4489c0e3\" (UID: \"d63db659-8460-4259-9396-80cd4489c0e3\") " Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.477806 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs" (OuterVolumeSpecName: "logs") pod "d63db659-8460-4259-9396-80cd4489c0e3" (UID: "d63db659-8460-4259-9396-80cd4489c0e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.481334 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm" (OuterVolumeSpecName: "kube-api-access-nxvmm") pod "d63db659-8460-4259-9396-80cd4489c0e3" (UID: "d63db659-8460-4259-9396-80cd4489c0e3"). InnerVolumeSpecName "kube-api-access-nxvmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.485875 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.508515 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data" (OuterVolumeSpecName: "config-data") pod "d63db659-8460-4259-9396-80cd4489c0e3" (UID: "d63db659-8460-4259-9396-80cd4489c0e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.512072 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d63db659-8460-4259-9396-80cd4489c0e3" (UID: "d63db659-8460-4259-9396-80cd4489c0e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.620454 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxvmm\" (UniqueName: \"kubernetes.io/projected/d63db659-8460-4259-9396-80cd4489c0e3-kube-api-access-nxvmm\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.620481 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d63db659-8460-4259-9396-80cd4489c0e3-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.620494 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.620505 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d63db659-8460-4259-9396-80cd4489c0e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:18 crc kubenswrapper[5018]: I1014 07:11:18.636651 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa6aaea-328b-4b70-ada7-e88dfcd098c3" path="/var/lib/kubelet/pods/3aa6aaea-328b-4b70-ada7-e88dfcd098c3/volumes" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.022963 5018 generic.go:334] "Generic (PLEG): container finished" podID="d63db659-8460-4259-9396-80cd4489c0e3" containerID="6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba" exitCode=0 Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.023026 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerDied","Data":"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba"} Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.023093 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d63db659-8460-4259-9396-80cd4489c0e3","Type":"ContainerDied","Data":"2c0bffcb10ed087a75fdb7965b076e0b7b2b5a0c3f9b5e51b5df6c9f40052c91"} Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.023115 5018 scope.go:117] "RemoveContainer" containerID="6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.023128 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.055825 5018 scope.go:117] "RemoveContainer" containerID="44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.060670 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.068296 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.077899 5018 scope.go:117] "RemoveContainer" containerID="6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba" Oct 14 07:11:19 crc kubenswrapper[5018]: E1014 07:11:19.079006 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba\": container with ID starting with 6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba not found: ID does not exist" containerID="6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.079051 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba"} err="failed to get container status \"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba\": rpc error: code = NotFound desc = could not find container \"6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba\": container with ID starting with 6f5715138b7ba195ba2799fc5220885a9282e54ae37068a872d474932caed5ba not found: ID does not exist" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.079081 5018 scope.go:117] "RemoveContainer" containerID="44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72" Oct 14 07:11:19 crc kubenswrapper[5018]: E1014 07:11:19.079662 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72\": container with ID starting with 44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72 not found: ID does not exist" containerID="44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.079695 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72"} err="failed to get container status \"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72\": rpc error: code = NotFound desc = could not find container \"44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72\": container with ID starting with 44ea91fc2559c049ade702c2cb3166bc49c70ac70064815dd585c4ab8486ff72 not found: ID does not exist" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.080953 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: E1014 07:11:19.081415 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-log" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.081442 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-log" Oct 14 07:11:19 crc kubenswrapper[5018]: E1014 07:11:19.081451 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-api" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.081458 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-api" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.081659 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-api" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.081688 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63db659-8460-4259-9396-80cd4489c0e3" containerName="nova-api-log" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.082656 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.087421 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.087669 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.087771 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.096708 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.136963 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.137026 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.137067 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.137095 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.137147 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-952pw\" (UniqueName: \"kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.137209 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.186340 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: W1014 07:11:19.194769 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d9387da_03d5_46e4_b442_aecb5fd1e766.slice/crio-2e29f4391ae9949b53ecb8add8a020c15b469a585747b89932d114a30311cbb0 WatchSource:0}: Error finding container 2e29f4391ae9949b53ecb8add8a020c15b469a585747b89932d114a30311cbb0: Status 404 returned error can't find the container with id 2e29f4391ae9949b53ecb8add8a020c15b469a585747b89932d114a30311cbb0 Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240522 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240584 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240612 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240651 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240677 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-952pw\" (UniqueName: \"kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.240723 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.242139 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.246115 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.246159 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.246798 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.247202 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.258512 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-952pw\" (UniqueName: \"kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw\") pod \"nova-api-0\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.445711 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:19 crc kubenswrapper[5018]: I1014 07:11:19.898950 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:19 crc kubenswrapper[5018]: W1014 07:11:19.910347 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b89f1e0_435d_4721_8937_b59d602d223a.slice/crio-d0612e578ef8a5928fcc8fa67ab4561c867bfb14b157b13583b28c05e6663573 WatchSource:0}: Error finding container d0612e578ef8a5928fcc8fa67ab4561c867bfb14b157b13583b28c05e6663573: Status 404 returned error can't find the container with id d0612e578ef8a5928fcc8fa67ab4561c867bfb14b157b13583b28c05e6663573 Oct 14 07:11:20 crc kubenswrapper[5018]: I1014 07:11:20.046909 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerStarted","Data":"d0612e578ef8a5928fcc8fa67ab4561c867bfb14b157b13583b28c05e6663573"} Oct 14 07:11:20 crc kubenswrapper[5018]: I1014 07:11:20.049934 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerStarted","Data":"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698"} Oct 14 07:11:20 crc kubenswrapper[5018]: I1014 07:11:20.050000 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerStarted","Data":"2e29f4391ae9949b53ecb8add8a020c15b469a585747b89932d114a30311cbb0"} Oct 14 07:11:20 crc kubenswrapper[5018]: I1014 07:11:20.619246 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63db659-8460-4259-9396-80cd4489c0e3" path="/var/lib/kubelet/pods/d63db659-8460-4259-9396-80cd4489c0e3/volumes" Oct 14 07:11:20 crc kubenswrapper[5018]: I1014 07:11:20.646794 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:21 crc kubenswrapper[5018]: I1014 07:11:21.062820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerStarted","Data":"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6"} Oct 14 07:11:21 crc kubenswrapper[5018]: I1014 07:11:21.063558 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerStarted","Data":"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850"} Oct 14 07:11:21 crc kubenswrapper[5018]: I1014 07:11:21.066958 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerStarted","Data":"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29"} Oct 14 07:11:21 crc kubenswrapper[5018]: I1014 07:11:21.092755 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.092727798 podStartE2EDuration="2.092727798s" podCreationTimestamp="2025-10-14 07:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:21.084731012 +0000 UTC m=+1297.668777639" watchObservedRunningTime="2025-10-14 07:11:21.092727798 +0000 UTC m=+1297.676774455" Oct 14 07:11:22 crc kubenswrapper[5018]: I1014 07:11:22.087521 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerStarted","Data":"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6"} Oct 14 07:11:22 crc kubenswrapper[5018]: I1014 07:11:22.396930 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:11:22 crc kubenswrapper[5018]: I1014 07:11:22.476816 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:11:22 crc kubenswrapper[5018]: I1014 07:11:22.477222 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="dnsmasq-dns" containerID="cri-o://1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64" gracePeriod=10 Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.041677 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.106043 5018 generic.go:334] "Generic (PLEG): container finished" podID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerID="1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64" exitCode=0 Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.106097 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" event={"ID":"f720b934-8cb2-4e78-8060-2fc0d23230fa","Type":"ContainerDied","Data":"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64"} Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.106127 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" event={"ID":"f720b934-8cb2-4e78-8060-2fc0d23230fa","Type":"ContainerDied","Data":"bfbe18bd917131d9b9a2cca3724d880223e7e3a46b80c1c4ecdacbd2a33d68eb"} Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.106148 5018 scope.go:117] "RemoveContainer" containerID="1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.106319 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d74749bf5-9rn9c" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.141134 5018 scope.go:117] "RemoveContainer" containerID="99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.164002 5018 scope.go:117] "RemoveContainer" containerID="1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64" Oct 14 07:11:23 crc kubenswrapper[5018]: E1014 07:11:23.164395 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64\": container with ID starting with 1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64 not found: ID does not exist" containerID="1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.164451 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64"} err="failed to get container status \"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64\": rpc error: code = NotFound desc = could not find container \"1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64\": container with ID starting with 1abb3a7a085f5f5a5efd43a1ba9eb4de67c449ae139cb41978a7ee898184de64 not found: ID does not exist" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.164484 5018 scope.go:117] "RemoveContainer" containerID="99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83" Oct 14 07:11:23 crc kubenswrapper[5018]: E1014 07:11:23.164749 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83\": container with ID starting with 99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83 not found: ID does not exist" containerID="99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.164780 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83"} err="failed to get container status \"99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83\": rpc error: code = NotFound desc = could not find container \"99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83\": container with ID starting with 99d9ba42644e4d0dad1ce42072dba7d8ef5126c4772d06336df77dd184061c83 not found: ID does not exist" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235542 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235634 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235841 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235877 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235909 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.235947 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zd7t\" (UniqueName: \"kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t\") pod \"f720b934-8cb2-4e78-8060-2fc0d23230fa\" (UID: \"f720b934-8cb2-4e78-8060-2fc0d23230fa\") " Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.243913 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t" (OuterVolumeSpecName: "kube-api-access-8zd7t") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "kube-api-access-8zd7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.301529 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.304533 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.304884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config" (OuterVolumeSpecName: "config") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.305462 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.328917 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f720b934-8cb2-4e78-8060-2fc0d23230fa" (UID: "f720b934-8cb2-4e78-8060-2fc0d23230fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338209 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338240 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338250 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338259 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338270 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f720b934-8cb2-4e78-8060-2fc0d23230fa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.338279 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zd7t\" (UniqueName: \"kubernetes.io/projected/f720b934-8cb2-4e78-8060-2fc0d23230fa-kube-api-access-8zd7t\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.440792 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:11:23 crc kubenswrapper[5018]: I1014 07:11:23.448912 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d74749bf5-9rn9c"] Oct 14 07:11:24 crc kubenswrapper[5018]: I1014 07:11:24.122738 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerStarted","Data":"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2"} Oct 14 07:11:24 crc kubenswrapper[5018]: I1014 07:11:24.123276 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:11:24 crc kubenswrapper[5018]: I1014 07:11:24.164223 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.038587722 podStartE2EDuration="6.164209954s" podCreationTimestamp="2025-10-14 07:11:18 +0000 UTC" firstStartedPulling="2025-10-14 07:11:19.196589309 +0000 UTC m=+1295.780635936" lastFinishedPulling="2025-10-14 07:11:23.322211541 +0000 UTC m=+1299.906258168" observedRunningTime="2025-10-14 07:11:24.159746747 +0000 UTC m=+1300.743793374" watchObservedRunningTime="2025-10-14 07:11:24.164209954 +0000 UTC m=+1300.748256581" Oct 14 07:11:24 crc kubenswrapper[5018]: I1014 07:11:24.626874 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" path="/var/lib/kubelet/pods/f720b934-8cb2-4e78-8060-2fc0d23230fa/volumes" Oct 14 07:11:25 crc kubenswrapper[5018]: I1014 07:11:25.647061 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:25 crc kubenswrapper[5018]: I1014 07:11:25.676149 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.181070 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.428247 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dsfhk"] Oct 14 07:11:26 crc kubenswrapper[5018]: E1014 07:11:26.428788 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="init" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.428811 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="init" Oct 14 07:11:26 crc kubenswrapper[5018]: E1014 07:11:26.428848 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="dnsmasq-dns" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.428857 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="dnsmasq-dns" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.429152 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f720b934-8cb2-4e78-8060-2fc0d23230fa" containerName="dnsmasq-dns" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.430053 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.431906 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.435409 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsfhk"] Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.444247 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.509331 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.509634 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.509658 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.509700 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttqxh\" (UniqueName: \"kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.611118 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.611180 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.611205 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.611242 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttqxh\" (UniqueName: \"kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.618809 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.620103 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.630457 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.632371 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttqxh\" (UniqueName: \"kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh\") pod \"nova-cell1-cell-mapping-dsfhk\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:26 crc kubenswrapper[5018]: I1014 07:11:26.757338 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:27 crc kubenswrapper[5018]: I1014 07:11:27.252052 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsfhk"] Oct 14 07:11:27 crc kubenswrapper[5018]: W1014 07:11:27.253413 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11b2e5f4_648d_47f5_a6c3_3a6621026889.slice/crio-07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1 WatchSource:0}: Error finding container 07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1: Status 404 returned error can't find the container with id 07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1 Oct 14 07:11:28 crc kubenswrapper[5018]: I1014 07:11:28.187107 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsfhk" event={"ID":"11b2e5f4-648d-47f5-a6c3-3a6621026889","Type":"ContainerStarted","Data":"116ecd9653cec8dd9db0afce5cfcb53c0fbe1151c252757153b99e432a73a32f"} Oct 14 07:11:28 crc kubenswrapper[5018]: I1014 07:11:28.187491 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsfhk" event={"ID":"11b2e5f4-648d-47f5-a6c3-3a6621026889","Type":"ContainerStarted","Data":"07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1"} Oct 14 07:11:28 crc kubenswrapper[5018]: I1014 07:11:28.210392 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dsfhk" podStartSLOduration=2.210377485 podStartE2EDuration="2.210377485s" podCreationTimestamp="2025-10-14 07:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:28.204629192 +0000 UTC m=+1304.788675819" watchObservedRunningTime="2025-10-14 07:11:28.210377485 +0000 UTC m=+1304.794424112" Oct 14 07:11:29 crc kubenswrapper[5018]: I1014 07:11:29.446615 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:29 crc kubenswrapper[5018]: I1014 07:11:29.447998 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:30 crc kubenswrapper[5018]: I1014 07:11:30.494825 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:30 crc kubenswrapper[5018]: I1014 07:11:30.494973 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:32 crc kubenswrapper[5018]: I1014 07:11:32.240936 5018 generic.go:334] "Generic (PLEG): container finished" podID="11b2e5f4-648d-47f5-a6c3-3a6621026889" containerID="116ecd9653cec8dd9db0afce5cfcb53c0fbe1151c252757153b99e432a73a32f" exitCode=0 Oct 14 07:11:32 crc kubenswrapper[5018]: I1014 07:11:32.241046 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsfhk" event={"ID":"11b2e5f4-648d-47f5-a6c3-3a6621026889","Type":"ContainerDied","Data":"116ecd9653cec8dd9db0afce5cfcb53c0fbe1151c252757153b99e432a73a32f"} Oct 14 07:11:32 crc kubenswrapper[5018]: I1014 07:11:32.463241 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:11:32 crc kubenswrapper[5018]: I1014 07:11:32.463329 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.791694 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.971474 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle\") pod \"11b2e5f4-648d-47f5-a6c3-3a6621026889\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.971834 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts\") pod \"11b2e5f4-648d-47f5-a6c3-3a6621026889\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.972063 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data\") pod \"11b2e5f4-648d-47f5-a6c3-3a6621026889\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.972178 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttqxh\" (UniqueName: \"kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh\") pod \"11b2e5f4-648d-47f5-a6c3-3a6621026889\" (UID: \"11b2e5f4-648d-47f5-a6c3-3a6621026889\") " Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.982994 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh" (OuterVolumeSpecName: "kube-api-access-ttqxh") pod "11b2e5f4-648d-47f5-a6c3-3a6621026889" (UID: "11b2e5f4-648d-47f5-a6c3-3a6621026889"). InnerVolumeSpecName "kube-api-access-ttqxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:33 crc kubenswrapper[5018]: I1014 07:11:33.983026 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts" (OuterVolumeSpecName: "scripts") pod "11b2e5f4-648d-47f5-a6c3-3a6621026889" (UID: "11b2e5f4-648d-47f5-a6c3-3a6621026889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.021564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11b2e5f4-648d-47f5-a6c3-3a6621026889" (UID: "11b2e5f4-648d-47f5-a6c3-3a6621026889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.024375 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data" (OuterVolumeSpecName: "config-data") pod "11b2e5f4-648d-47f5-a6c3-3a6621026889" (UID: "11b2e5f4-648d-47f5-a6c3-3a6621026889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.075294 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.075352 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttqxh\" (UniqueName: \"kubernetes.io/projected/11b2e5f4-648d-47f5-a6c3-3a6621026889-kube-api-access-ttqxh\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.075374 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.075477 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11b2e5f4-648d-47f5-a6c3-3a6621026889-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.268206 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsfhk" event={"ID":"11b2e5f4-648d-47f5-a6c3-3a6621026889","Type":"ContainerDied","Data":"07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1"} Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.268530 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07efe5a9bb97f22dc55f360373dae33b64aa33064b6a586f9be8032d37bd76c1" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.268312 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsfhk" Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.474557 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.474915 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-log" containerID="cri-o://16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850" gracePeriod=30 Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.475119 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-api" containerID="cri-o://5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6" gracePeriod=30 Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.500416 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.500856 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" containerName="nova-scheduler-scheduler" containerID="cri-o://ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae" gracePeriod=30 Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.519162 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.519611 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" containerID="cri-o://8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481" gracePeriod=30 Oct 14 07:11:34 crc kubenswrapper[5018]: I1014 07:11:34.519984 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" containerID="cri-o://dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d" gracePeriod=30 Oct 14 07:11:35 crc kubenswrapper[5018]: I1014 07:11:35.279348 5018 generic.go:334] "Generic (PLEG): container finished" podID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerID="8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481" exitCode=143 Oct 14 07:11:35 crc kubenswrapper[5018]: I1014 07:11:35.279459 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerDied","Data":"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481"} Oct 14 07:11:35 crc kubenswrapper[5018]: I1014 07:11:35.283126 5018 generic.go:334] "Generic (PLEG): container finished" podID="1b89f1e0-435d-4721-8937-b59d602d223a" containerID="16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850" exitCode=143 Oct 14 07:11:35 crc kubenswrapper[5018]: I1014 07:11:35.283177 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerDied","Data":"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850"} Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.034668 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.211428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data\") pod \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.211852 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg9d7\" (UniqueName: \"kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7\") pod \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.211886 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle\") pod \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\" (UID: \"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8\") " Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.226593 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7" (OuterVolumeSpecName: "kube-api-access-sg9d7") pod "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" (UID: "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8"). InnerVolumeSpecName "kube-api-access-sg9d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.251118 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" (UID: "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.257192 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data" (OuterVolumeSpecName: "config-data") pod "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" (UID: "a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.295739 5018 generic.go:334] "Generic (PLEG): container finished" podID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" containerID="ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae" exitCode=0 Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.295779 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8","Type":"ContainerDied","Data":"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae"} Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.295805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8","Type":"ContainerDied","Data":"85229297bddc723820b9e066fc49cee113ad2efcc452d13cdce8e927b8c6d896"} Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.295820 5018 scope.go:117] "RemoveContainer" containerID="ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.295916 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.314207 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.314237 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg9d7\" (UniqueName: \"kubernetes.io/projected/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-kube-api-access-sg9d7\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.314249 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.379201 5018 scope.go:117] "RemoveContainer" containerID="ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae" Oct 14 07:11:36 crc kubenswrapper[5018]: E1014 07:11:36.379567 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae\": container with ID starting with ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae not found: ID does not exist" containerID="ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.379599 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae"} err="failed to get container status \"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae\": rpc error: code = NotFound desc = could not find container \"ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae\": container with ID starting with ff5f1628f228c1fb20b886d84716ab82de7e7c3c13aa0f0410ac31730ab7aeae not found: ID does not exist" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.381901 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.396173 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.411721 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:36 crc kubenswrapper[5018]: E1014 07:11:36.412170 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" containerName="nova-scheduler-scheduler" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.412188 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" containerName="nova-scheduler-scheduler" Oct 14 07:11:36 crc kubenswrapper[5018]: E1014 07:11:36.412216 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b2e5f4-648d-47f5-a6c3-3a6621026889" containerName="nova-manage" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.412224 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b2e5f4-648d-47f5-a6c3-3a6621026889" containerName="nova-manage" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.412426 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" containerName="nova-scheduler-scheduler" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.412444 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b2e5f4-648d-47f5-a6c3-3a6621026889" containerName="nova-manage" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.413089 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.419358 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.422510 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.518564 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbkvb\" (UniqueName: \"kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.518843 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.518961 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.622169 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbkvb\" (UniqueName: \"kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.622382 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.622458 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.632947 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.640084 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.649736 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8" path="/var/lib/kubelet/pods/a42b8d6a-9b7f-4457-b8db-32cf7d1b8bf8/volumes" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.666600 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbkvb\" (UniqueName: \"kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb\") pod \"nova-scheduler-0\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " pod="openstack/nova-scheduler-0" Oct 14 07:11:36 crc kubenswrapper[5018]: I1014 07:11:36.727996 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:11:37 crc kubenswrapper[5018]: W1014 07:11:37.262818 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod467829ec_f0b0_490c_80fb_25dc08df07b4.slice/crio-0647c750323f18caa6a330d640f5a72168c770d7b54ec4109677bf8f679fc69c WatchSource:0}: Error finding container 0647c750323f18caa6a330d640f5a72168c770d7b54ec4109677bf8f679fc69c: Status 404 returned error can't find the container with id 0647c750323f18caa6a330d640f5a72168c770d7b54ec4109677bf8f679fc69c Oct 14 07:11:37 crc kubenswrapper[5018]: I1014 07:11:37.264541 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:11:37 crc kubenswrapper[5018]: I1014 07:11:37.310592 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"467829ec-f0b0-490c-80fb-25dc08df07b4","Type":"ContainerStarted","Data":"0647c750323f18caa6a330d640f5a72168c770d7b54ec4109677bf8f679fc69c"} Oct 14 07:11:37 crc kubenswrapper[5018]: I1014 07:11:37.693097 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:45302->10.217.0.194:8775: read: connection reset by peer" Oct 14 07:11:37 crc kubenswrapper[5018]: I1014 07:11:37.693096 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:45316->10.217.0.194:8775: read: connection reset by peer" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.147571 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.173523 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259270 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259332 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle\") pod \"8c73248c-c7c2-4740-887a-52abecf1b6f8\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259388 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-952pw\" (UniqueName: \"kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259452 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq9fg\" (UniqueName: \"kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg\") pod \"8c73248c-c7c2-4740-887a-52abecf1b6f8\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259524 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs\") pod \"8c73248c-c7c2-4740-887a-52abecf1b6f8\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259559 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs\") pod \"8c73248c-c7c2-4740-887a-52abecf1b6f8\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259598 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259657 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259743 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data\") pod \"8c73248c-c7c2-4740-887a-52abecf1b6f8\" (UID: \"8c73248c-c7c2-4740-887a-52abecf1b6f8\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259773 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.259843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data\") pod \"1b89f1e0-435d-4721-8937-b59d602d223a\" (UID: \"1b89f1e0-435d-4721-8937-b59d602d223a\") " Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.260912 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs" (OuterVolumeSpecName: "logs") pod "8c73248c-c7c2-4740-887a-52abecf1b6f8" (UID: "8c73248c-c7c2-4740-887a-52abecf1b6f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.261311 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs" (OuterVolumeSpecName: "logs") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.287482 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw" (OuterVolumeSpecName: "kube-api-access-952pw") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "kube-api-access-952pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.287948 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg" (OuterVolumeSpecName: "kube-api-access-fq9fg") pod "8c73248c-c7c2-4740-887a-52abecf1b6f8" (UID: "8c73248c-c7c2-4740-887a-52abecf1b6f8"). InnerVolumeSpecName "kube-api-access-fq9fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.306626 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.324490 5018 generic.go:334] "Generic (PLEG): container finished" podID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerID="dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d" exitCode=0 Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.324566 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerDied","Data":"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d"} Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.324599 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c73248c-c7c2-4740-887a-52abecf1b6f8","Type":"ContainerDied","Data":"93a501a273d9d68cd6145592e5c387fb43334ddb7fe7deb007bf4d7bc950e7f1"} Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.324635 5018 scope.go:117] "RemoveContainer" containerID="dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.324775 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.331042 5018 generic.go:334] "Generic (PLEG): container finished" podID="1b89f1e0-435d-4721-8937-b59d602d223a" containerID="5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6" exitCode=0 Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.331106 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.331130 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerDied","Data":"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6"} Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.331201 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b89f1e0-435d-4721-8937-b59d602d223a","Type":"ContainerDied","Data":"d0612e578ef8a5928fcc8fa67ab4561c867bfb14b157b13583b28c05e6663573"} Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.332744 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"467829ec-f0b0-490c-80fb-25dc08df07b4","Type":"ContainerStarted","Data":"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983"} Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.333853 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data" (OuterVolumeSpecName: "config-data") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363123 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq9fg\" (UniqueName: \"kubernetes.io/projected/8c73248c-c7c2-4740-887a-52abecf1b6f8-kube-api-access-fq9fg\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363197 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73248c-c7c2-4740-887a-52abecf1b6f8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363208 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b89f1e0-435d-4721-8937-b59d602d223a-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363216 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363226 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.363235 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-952pw\" (UniqueName: \"kubernetes.io/projected/1b89f1e0-435d-4721-8937-b59d602d223a-kube-api-access-952pw\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.366360 5018 scope.go:117] "RemoveContainer" containerID="8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.373594 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data" (OuterVolumeSpecName: "config-data") pod "8c73248c-c7c2-4740-887a-52abecf1b6f8" (UID: "8c73248c-c7c2-4740-887a-52abecf1b6f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.375160 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c73248c-c7c2-4740-887a-52abecf1b6f8" (UID: "8c73248c-c7c2-4740-887a-52abecf1b6f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.377986 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.380279 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8c73248c-c7c2-4740-887a-52abecf1b6f8" (UID: "8c73248c-c7c2-4740-887a-52abecf1b6f8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.388492 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1b89f1e0-435d-4721-8937-b59d602d223a" (UID: "1b89f1e0-435d-4721-8937-b59d602d223a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.390516 5018 scope.go:117] "RemoveContainer" containerID="dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.390959 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d\": container with ID starting with dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d not found: ID does not exist" containerID="dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.390989 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d"} err="failed to get container status \"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d\": rpc error: code = NotFound desc = could not find container \"dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d\": container with ID starting with dfc73d6958a5865382b08a9292fa0ae50dc64b4bbebcfb56ab337a5f8c08317d not found: ID does not exist" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.391010 5018 scope.go:117] "RemoveContainer" containerID="8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.391279 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481\": container with ID starting with 8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481 not found: ID does not exist" containerID="8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.391309 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481"} err="failed to get container status \"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481\": rpc error: code = NotFound desc = could not find container \"8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481\": container with ID starting with 8387acfc19a8d09170a40127a0d803f8b75f36d7df158889baca56032ef8c481 not found: ID does not exist" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.391328 5018 scope.go:117] "RemoveContainer" containerID="5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.406761 5018 scope.go:117] "RemoveContainer" containerID="16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.423178 5018 scope.go:117] "RemoveContainer" containerID="5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.423586 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6\": container with ID starting with 5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6 not found: ID does not exist" containerID="5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.423675 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6"} err="failed to get container status \"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6\": rpc error: code = NotFound desc = could not find container \"5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6\": container with ID starting with 5818202fbd9dc0e21205c631c056c2c6daec7d8b67721ea1fb9b3f50c8acb2d6 not found: ID does not exist" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.423762 5018 scope.go:117] "RemoveContainer" containerID="16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.424097 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850\": container with ID starting with 16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850 not found: ID does not exist" containerID="16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.424149 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850"} err="failed to get container status \"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850\": rpc error: code = NotFound desc = could not find container \"16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850\": container with ID starting with 16045289a6ef4a86515d459aac6b0a6d0ea35ca9e1e9146aeef86bf2c9d1e850 not found: ID does not exist" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.465438 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.465476 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.465489 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.465501 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b89f1e0-435d-4721-8937-b59d602d223a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.465512 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73248c-c7c2-4740-887a-52abecf1b6f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.653870 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.65385391 podStartE2EDuration="2.65385391s" podCreationTimestamp="2025-10-14 07:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:38.361880125 +0000 UTC m=+1314.945926752" watchObservedRunningTime="2025-10-14 07:11:38.65385391 +0000 UTC m=+1315.237900537" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.662432 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.673370 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.691515 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.692215 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692242 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.692277 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-api" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692288 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-api" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.692307 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692315 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" Oct 14 07:11:38 crc kubenswrapper[5018]: E1014 07:11:38.692327 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-log" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692335 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-log" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692574 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-log" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692591 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" containerName="nova-api-api" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692606 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-metadata" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.692674 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" containerName="nova-metadata-log" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.693847 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.701941 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.702477 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.702910 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.711165 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.764189 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771045 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771181 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pbm4\" (UniqueName: \"kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771241 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771325 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771408 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771467 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.771526 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.795853 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.797729 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.799915 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.800021 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.805521 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.872735 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.872782 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f66l\" (UniqueName: \"kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.872833 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.872874 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pbm4\" (UniqueName: \"kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873440 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.872892 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873529 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873591 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873668 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873720 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.873765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.877410 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.878090 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.878159 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.888012 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.892939 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pbm4\" (UniqueName: \"kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4\") pod \"nova-api-0\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " pod="openstack/nova-api-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.975520 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.975649 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.975726 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f66l\" (UniqueName: \"kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.975868 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.975906 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.976028 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.979534 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.980369 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.983493 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:38 crc kubenswrapper[5018]: I1014 07:11:38.999150 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f66l\" (UniqueName: \"kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l\") pod \"nova-metadata-0\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " pod="openstack/nova-metadata-0" Oct 14 07:11:39 crc kubenswrapper[5018]: I1014 07:11:39.078250 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:11:39 crc kubenswrapper[5018]: I1014 07:11:39.126148 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:11:39 crc kubenswrapper[5018]: W1014 07:11:39.584955 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6a69b66_5d0d_410b_9d2e_6179c8079077.slice/crio-6b777a2685bad24921e479b5e2015642fb370e2734cce4591099a6578f181128 WatchSource:0}: Error finding container 6b777a2685bad24921e479b5e2015642fb370e2734cce4591099a6578f181128: Status 404 returned error can't find the container with id 6b777a2685bad24921e479b5e2015642fb370e2734cce4591099a6578f181128 Oct 14 07:11:39 crc kubenswrapper[5018]: I1014 07:11:39.593605 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:11:39 crc kubenswrapper[5018]: I1014 07:11:39.603336 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.376916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerStarted","Data":"6c69169eb51092731bb9a425eec4f189573257f2c5056b90bbbf411e5e89d260"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.377280 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerStarted","Data":"f7f8ce51b32d6ce79cf26bf72b40b8a1faaf32d35ad244f801fb5938a444418c"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.377298 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerStarted","Data":"6b777a2685bad24921e479b5e2015642fb370e2734cce4591099a6578f181128"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.382448 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerStarted","Data":"620014af1412ba75636f9ca794cb656dd537c7d697efc2facc190d885dbfce9b"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.382492 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerStarted","Data":"27b9f6c601d30ddeef6ae8d67f22e5d31f0b0856e67d3120476a411a44292642"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.382507 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerStarted","Data":"5bebb8f7a6fb0607c3475c0663ea8e8f358eea574d7a1f95b900d716594ff4aa"} Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.404913 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4048975759999998 podStartE2EDuration="2.404897576s" podCreationTimestamp="2025-10-14 07:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:40.397229229 +0000 UTC m=+1316.981275856" watchObservedRunningTime="2025-10-14 07:11:40.404897576 +0000 UTC m=+1316.988944203" Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.424753 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.424737238 podStartE2EDuration="2.424737238s" podCreationTimestamp="2025-10-14 07:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:11:40.415406544 +0000 UTC m=+1316.999453181" watchObservedRunningTime="2025-10-14 07:11:40.424737238 +0000 UTC m=+1317.008783855" Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.624349 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b89f1e0-435d-4721-8937-b59d602d223a" path="/var/lib/kubelet/pods/1b89f1e0-435d-4721-8937-b59d602d223a/volumes" Oct 14 07:11:40 crc kubenswrapper[5018]: I1014 07:11:40.625501 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c73248c-c7c2-4740-887a-52abecf1b6f8" path="/var/lib/kubelet/pods/8c73248c-c7c2-4740-887a-52abecf1b6f8/volumes" Oct 14 07:11:41 crc kubenswrapper[5018]: I1014 07:11:41.729313 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:11:44 crc kubenswrapper[5018]: I1014 07:11:44.127371 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:11:44 crc kubenswrapper[5018]: I1014 07:11:44.127817 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:11:46 crc kubenswrapper[5018]: I1014 07:11:46.729119 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 07:11:46 crc kubenswrapper[5018]: I1014 07:11:46.789490 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 07:11:47 crc kubenswrapper[5018]: I1014 07:11:47.499984 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 07:11:48 crc kubenswrapper[5018]: I1014 07:11:48.498708 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 07:11:49 crc kubenswrapper[5018]: I1014 07:11:49.080743 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:49 crc kubenswrapper[5018]: I1014 07:11:49.080823 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:11:49 crc kubenswrapper[5018]: I1014 07:11:49.126533 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:11:49 crc kubenswrapper[5018]: I1014 07:11:49.126597 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:11:50 crc kubenswrapper[5018]: I1014 07:11:50.090754 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:50 crc kubenswrapper[5018]: I1014 07:11:50.090764 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:50 crc kubenswrapper[5018]: I1014 07:11:50.151844 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:50 crc kubenswrapper[5018]: I1014 07:11:50.151948 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.092174 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.093003 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.093805 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.093951 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.101551 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.103240 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.139990 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.141800 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.198566 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:11:59 crc kubenswrapper[5018]: I1014 07:11:59.614186 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:12:02 crc kubenswrapper[5018]: I1014 07:12:02.463183 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:12:02 crc kubenswrapper[5018]: I1014 07:12:02.463710 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.591498 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.592125 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" containerName="openstackclient" containerID="cri-o://672ba654264cd63e079885cb542e177121988e36e45b223e1019d3465e9c3dcc" gracePeriod=2 Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.606415 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.817005 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.836471 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:21 crc kubenswrapper[5018]: E1014 07:12:21.836897 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" containerName="openstackclient" Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.836911 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" containerName="openstackclient" Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.837086 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" containerName="openstackclient" Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.839640 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.862604 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.941928 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.948035 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="openstack-network-exporter" containerID="cri-o://9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a" gracePeriod=30 Oct 14 07:12:21 crc kubenswrapper[5018]: I1014 07:12:21.948269 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" containerID="cri-o://93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" gracePeriod=30 Oct 14 07:12:21 crc kubenswrapper[5018]: E1014 07:12:21.949486 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:21 crc kubenswrapper[5018]: E1014 07:12:21.949519 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data podName:1dae2a92-e082-4271-b517-2f9219a07f2b nodeName:}" failed. No retries permitted until 2025-10-14 07:12:22.449506818 +0000 UTC m=+1359.033553445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data") pod "rabbitmq-cell1-server-0" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.031702 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.038356 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.042746 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcw6w\" (UniqueName: \"kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w\") pod \"cinder5769-account-delete-pwqc4\" (UID: \"6719ab90-8279-40bf-9544-b35118039b0e\") " pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.088452 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.148373 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcw6w\" (UniqueName: \"kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w\") pod \"cinder5769-account-delete-pwqc4\" (UID: \"6719ab90-8279-40bf-9544-b35118039b0e\") " pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.148519 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btjxb\" (UniqueName: \"kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb\") pod \"glancedba2-account-delete-fhm4g\" (UID: \"17ef95f6-1b6d-459a-b331-b114fb6205fb\") " pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.196109 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.197240 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.198588 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcw6w\" (UniqueName: \"kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w\") pod \"cinder5769-account-delete-pwqc4\" (UID: \"6719ab90-8279-40bf-9544-b35118039b0e\") " pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.239148 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.250005 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btjxb\" (UniqueName: \"kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb\") pod \"glancedba2-account-delete-fhm4g\" (UID: \"17ef95f6-1b6d-459a-b331-b114fb6205fb\") " pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.352275 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnmlf\" (UniqueName: \"kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf\") pod \"placementd72f-account-delete-7nsjg\" (UID: \"d2c0a409-0cc4-49a4-a1ad-3de144f90e49\") " pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.357209 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btjxb\" (UniqueName: \"kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb\") pod \"glancedba2-account-delete-fhm4g\" (UID: \"17ef95f6-1b6d-459a-b331-b114fb6205fb\") " pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.395878 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.396432 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="openstack-network-exporter" containerID="cri-o://015528ec03a9ccae5d8175782d21df263addf1920d63d59b5c8dd6f6c5ecbd26" gracePeriod=300 Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.411766 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-2b2jk"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.454498 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnmlf\" (UniqueName: \"kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf\") pod \"placementd72f-account-delete-7nsjg\" (UID: \"d2c0a409-0cc4-49a4-a1ad-3de144f90e49\") " pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:22 crc kubenswrapper[5018]: E1014 07:12:22.454713 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:22 crc kubenswrapper[5018]: E1014 07:12:22.454767 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data podName:1dae2a92-e082-4271-b517-2f9219a07f2b nodeName:}" failed. No retries permitted until 2025-10-14 07:12:23.454750747 +0000 UTC m=+1360.038797374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data") pod "rabbitmq-cell1-server-0" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.455381 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-2b2jk"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.477383 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.478851 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.484263 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.496786 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.505228 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.540215 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.540556 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="openstack-network-exporter" containerID="cri-o://55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6" gracePeriod=300 Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.558893 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wskhd\" (UniqueName: \"kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd\") pod \"neutron169d-account-delete-gjr5f\" (UID: \"60a3ff62-8fcf-47aa-b080-b94042e137be\") " pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.600939 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnmlf\" (UniqueName: \"kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf\") pod \"placementd72f-account-delete-7nsjg\" (UID: \"d2c0a409-0cc4-49a4-a1ad-3de144f90e49\") " pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.601009 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-77v4n"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.663859 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wskhd\" (UniqueName: \"kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd\") pod \"neutron169d-account-delete-gjr5f\" (UID: \"60a3ff62-8fcf-47aa-b080-b94042e137be\") " pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.686878 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25622217-99df-4f81-a099-b17af99a7dd6" path="/var/lib/kubelet/pods/25622217-99df-4f81-a099-b17af99a7dd6/volumes" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.687524 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-77v4n"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.717110 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4nzkd"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.736914 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4nzkd"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.740587 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wskhd\" (UniqueName: \"kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd\") pod \"neutron169d-account-delete-gjr5f\" (UID: \"60a3ff62-8fcf-47aa-b080-b94042e137be\") " pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.747773 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.767395 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nx7zt"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.778439 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nx7zt"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.793515 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="ovsdbserver-nb" containerID="cri-o://ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856" gracePeriod=300 Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.853246 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.861090 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.862269 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.872210 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:22 crc kubenswrapper[5018]: E1014 07:12:22.874348 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:12:22 crc kubenswrapper[5018]: E1014 07:12:22.874404 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data podName:5ddcde8d-45fd-419b-8751-18db89bf729d nodeName:}" failed. No retries permitted until 2025-10-14 07:12:23.374381809 +0000 UTC m=+1359.958428426 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data") pod "rabbitmq-server-0" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d") : configmap "rabbitmq-config-data" not found Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.890890 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.980135 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq\") pod \"novaapif4bd-account-delete-vrgxk\" (UID: \"6e0a409e-4a1f-471e-a5d5-b01213e5c0df\") " pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.991316 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:12:22 crc kubenswrapper[5018]: I1014 07:12:22.991839 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-7pr99" podUID="0f737b49-f2ba-4a7c-9b07-d934d5122363" containerName="openstack-network-exporter" containerID="cri-o://3b16b9017596763bb0f70d7dd6f8aca7f1f4fbb0b4ed9d4a7005193cc2610d01" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:22.995707 5018 generic.go:334] "Generic (PLEG): container finished" podID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerID="015528ec03a9ccae5d8175782d21df263addf1920d63d59b5c8dd6f6c5ecbd26" exitCode=2 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:22.995779 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerDied","Data":"015528ec03a9ccae5d8175782d21df263addf1920d63d59b5c8dd6f6c5ecbd26"} Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.010068 5018 generic.go:334] "Generic (PLEG): container finished" podID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerID="9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a" exitCode=2 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.010414 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerDied","Data":"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a"} Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.023677 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="ovsdbserver-sb" containerID="cri-o://f1c52505d00c71d74b81b018f8807279db186987ee6fc9cf83ce290b691b50e9" gracePeriod=300 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.052355 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c2372cde-de88-46c2-abc8-0bd1a9e7b961/ovsdbserver-nb/0.log" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.052721 5018 generic.go:334] "Generic (PLEG): container finished" podID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerID="55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6" exitCode=2 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.052754 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerDied","Data":"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6"} Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.066513 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.088049 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq\") pod \"novaapif4bd-account-delete-vrgxk\" (UID: \"6e0a409e-4a1f-471e-a5d5-b01213e5c0df\") " pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.093673 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.120665 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pnqtc"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.146930 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pnqtc"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.168194 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq\") pod \"novaapif4bd-account-delete-vrgxk\" (UID: \"6e0a409e-4a1f-471e-a5d5-b01213e5c0df\") " pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.212709 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.228294 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.244306 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.275870 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.276103 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65bf758599-jh97x" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="dnsmasq-dns" containerID="cri-o://f563d6fe835cc693edf98b5acc197853d2d8d196613cf1c29f3ee82afbf51623" gracePeriod=10 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.306078 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmlbm\" (UniqueName: \"kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm\") pod \"novacell0fab3-account-delete-l4j7n\" (UID: \"f5d08361-7993-4757-afe3-1e9eeb5c1448\") " pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.315080 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.315643 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.363538 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.363783 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api-log" containerID="cri-o://a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.364925 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api" containerID="cri-o://b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.365747 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.390892 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.390936 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.408173 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmlbm\" (UniqueName: \"kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm\") pod \"novacell0fab3-account-delete-l4j7n\" (UID: \"f5d08361-7993-4757-afe3-1e9eeb5c1448\") " pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.408568 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.408610 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data podName:5ddcde8d-45fd-419b-8751-18db89bf729d nodeName:}" failed. No retries permitted until 2025-10-14 07:12:24.40859703 +0000 UTC m=+1360.992643657 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data") pod "rabbitmq-server-0" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d") : configmap "rabbitmq-config-data" not found Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.447929 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.448312 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="cinder-scheduler" containerID="cri-o://73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.459729 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="probe" containerID="cri-o://f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.461955 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmlbm\" (UniqueName: \"kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm\") pod \"novacell0fab3-account-delete-l4j7n\" (UID: \"f5d08361-7993-4757-afe3-1e9eeb5c1448\") " pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.513126 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:23 crc kubenswrapper[5018]: E1014 07:12:23.513201 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data podName:1dae2a92-e082-4271-b517-2f9219a07f2b nodeName:}" failed. No retries permitted until 2025-10-14 07:12:25.513178844 +0000 UTC m=+1362.097225471 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data") pod "rabbitmq-cell1-server-0" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.517103 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w4h5j"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.564108 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-w4h5j"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.606467 5018 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/nova-cell1-conductor-0" secret="" err="secret \"nova-nova-dockercfg-tnjd5\" not found" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.612669 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.613103 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-log" containerID="cri-o://53f8bbd9b14b8748fd6d68d48207701da31ebb8e7309185ebed668c22c0089ca" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.613498 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-httpd" containerID="cri-o://b96379567c4eaad728194e40c42a49fa57be113eff3ccca7f60a2b9c2df42bbb" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.635091 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.635361 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7649bf8b6b-5szh6" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-log" containerID="cri-o://d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.635768 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7649bf8b6b-5szh6" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-api" containerID="cri-o://6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.650552 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.650791 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75695fb6c7-7ztlf" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-api" containerID="cri-o://d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.652654 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75695fb6c7-7ztlf" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-httpd" containerID="cri-o://fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.679032 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.711676 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712144 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-server" containerID="cri-o://422e0f3e0acbeb85eabbf4a53967dd8d157d1f9364ac749656bd718ee3f5b1e7" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712525 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="swift-recon-cron" containerID="cri-o://f75caf8dbbd7c4743e32fa90f2af02023f5f583c5e27a58a4d2ae6216d5c09b9" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712578 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="rsync" containerID="cri-o://7464ef303afd6563d9b11c50e58be39b3a6ce6d6bc0c340a249ad132b5dfdcba" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712612 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-expirer" containerID="cri-o://969d5f8aadb42ee0593e5d86418c4fdb29ca61ce0f36d2a945eec3737e42b189" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712666 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-updater" containerID="cri-o://ec680e0a29290410e29c3d3731406e8216b29a7138c4bf28271899256ee732cc" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712697 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-auditor" containerID="cri-o://599bb15e72a850b4e7c1ebc86b49f02ba6953b1cd2083f0f540b8f3b91ff08c0" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712726 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-replicator" containerID="cri-o://963b3d72e03491fd9f8e960e77328e431e858c28e5183636d20edce5d6da90ec" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712763 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-server" containerID="cri-o://1c34c44d438a3d6af5f9816098723851b45124c4a7a2d7e1641ec047b86764eb" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712791 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-updater" containerID="cri-o://b05e853fcda282db718ede400b9daeb449fd5e7cd0b75409d593e3b167779576" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712819 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-auditor" containerID="cri-o://ebd9c6ab8467a65fedbce41b1ce78986ad397f0ca50597aaea953816e194054d" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712846 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-replicator" containerID="cri-o://80c27ba55cfa5a682b1e2072de8e447382019d0b79b38cf3f4de1b47c35a429d" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712875 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-server" containerID="cri-o://e58847c0dec122f46ec351ebf113ea26a71f6dd0002b59bf78203f30ea173dca" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712903 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-reaper" containerID="cri-o://ee71c8af72a00afcf6ab9a1af606dec3a2657101b03f635f39d3c3a416113250" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712935 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-auditor" containerID="cri-o://a9db9d72a3a3664bb31edfeb916708e025a2ea1bc87363e6693b7d4ae2caf69c" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.712961 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-replicator" containerID="cri-o://b83fc34a9d177c956e778ff75097967bb0fc448f9d4a508069b6032e5cd85729" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.720214 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.720413 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-log" containerID="cri-o://fb571e1ab7e3e5551c801254e6ef0537cf870849ca12640092089f0011af7c29" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.720544 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-httpd" containerID="cri-o://84742c9fbf39092ccf977df6e5a65995900f3bd0bbaf07ff5f7aed100b72b67f" gracePeriod=30 Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.771391 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhzhc"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.778390 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhzhc"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.792810 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsfhk"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.817705 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsfhk"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.831924 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2km9n"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.850943 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2km9n"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.864112 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.878790 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-dba2-account-create-8x996"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.891231 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-dba2-account-create-8x996"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.944250 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lrzn8"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.955638 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lrzn8"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.963803 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.979139 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0076-account-create-pdx2z"] Oct 14 07:12:23 crc kubenswrapper[5018]: I1014 07:12:23.992907 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0076-account-create-pdx2z"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.011861 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c2372cde-de88-46c2-abc8-0bd1a9e7b961/ovsdbserver-nb/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.011929 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.024163 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c4hlc"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.032562 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="rabbitmq" containerID="cri-o://db5a58f0d517ed36e8022372c1d0cbb0ac63dbc1e9c0b3d9627b4e9590b35c60" gracePeriod=604800 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.043688 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c4hlc"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.050754 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.051017 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-httpd" containerID="cri-o://e00c8d1ad5dfcec3691b831c54fc96d156f290741de76b782f82a372f3292ae3" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.051456 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-server" containerID="cri-o://2b02d899d0cced29508d36139ee74410a8e6281e06ca1fb7f832e68bbfcb9048" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.087093 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d72f-account-create-897bn"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.108144 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d72f-account-create-897bn"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.127428 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129128 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129205 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129234 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129259 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129303 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129447 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129494 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.129515 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqsx9\" (UniqueName: \"kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9\") pod \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\" (UID: \"c2372cde-de88-46c2-abc8-0bd1a9e7b961\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.137262 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config" (OuterVolumeSpecName: "config") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.137327 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts" (OuterVolumeSpecName: "scripts") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.137549 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.172279 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.172727 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9" (OuterVolumeSpecName: "kube-api-access-sqsx9") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "kube-api-access-sqsx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.245962 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.246042 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqsx9\" (UniqueName: \"kubernetes.io/projected/c2372cde-de88-46c2-abc8-0bd1a9e7b961-kube-api-access-sqsx9\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.246058 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.246069 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2372cde-de88-46c2-abc8-0bd1a9e7b961-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.246128 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256325 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="969d5f8aadb42ee0593e5d86418c4fdb29ca61ce0f36d2a945eec3737e42b189" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256379 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="ec680e0a29290410e29c3d3731406e8216b29a7138c4bf28271899256ee732cc" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256387 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="599bb15e72a850b4e7c1ebc86b49f02ba6953b1cd2083f0f540b8f3b91ff08c0" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256403 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="963b3d72e03491fd9f8e960e77328e431e858c28e5183636d20edce5d6da90ec" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256412 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="b05e853fcda282db718ede400b9daeb449fd5e7cd0b75409d593e3b167779576" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256418 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="ebd9c6ab8467a65fedbce41b1ce78986ad397f0ca50597aaea953816e194054d" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256425 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="80c27ba55cfa5a682b1e2072de8e447382019d0b79b38cf3f4de1b47c35a429d" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256476 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="ee71c8af72a00afcf6ab9a1af606dec3a2657101b03f635f39d3c3a416113250" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256486 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="a9db9d72a3a3664bb31edfeb916708e025a2ea1bc87363e6693b7d4ae2caf69c" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256493 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="b83fc34a9d177c956e778ff75097967bb0fc448f9d4a508069b6032e5cd85729" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256698 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fzlkk"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256721 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"969d5f8aadb42ee0593e5d86418c4fdb29ca61ce0f36d2a945eec3737e42b189"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256765 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"ec680e0a29290410e29c3d3731406e8216b29a7138c4bf28271899256ee732cc"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256795 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"599bb15e72a850b4e7c1ebc86b49f02ba6953b1cd2083f0f540b8f3b91ff08c0"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.256805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"963b3d72e03491fd9f8e960e77328e431e858c28e5183636d20edce5d6da90ec"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.268834 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"b05e853fcda282db718ede400b9daeb449fd5e7cd0b75409d593e3b167779576"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.268869 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"ebd9c6ab8467a65fedbce41b1ce78986ad397f0ca50597aaea953816e194054d"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.269969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"80c27ba55cfa5a682b1e2072de8e447382019d0b79b38cf3f4de1b47c35a429d"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.269990 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"ee71c8af72a00afcf6ab9a1af606dec3a2657101b03f635f39d3c3a416113250"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.270000 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"a9db9d72a3a3664bb31edfeb916708e025a2ea1bc87363e6693b7d4ae2caf69c"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.270011 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"b83fc34a9d177c956e778ff75097967bb0fc448f9d4a508069b6032e5cd85729"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.272966 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c2372cde-de88-46c2-abc8-0bd1a9e7b961/ovsdbserver-nb/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.273193 5018 generic.go:334] "Generic (PLEG): container finished" podID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerID="ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.273461 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.274370 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerDied","Data":"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.274419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c2372cde-de88-46c2-abc8-0bd1a9e7b961","Type":"ContainerDied","Data":"c453ceed07ab2e563923f41d67fe3f669d93551cadb48fc652de1444a662220e"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.274437 5018 scope.go:117] "RemoveContainer" containerID="55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.299704 5018 generic.go:334] "Generic (PLEG): container finished" podID="1719e0be-9beb-4f35-b449-2a90e51ec80c" containerID="672ba654264cd63e079885cb542e177121988e36e45b223e1019d3465e9c3dcc" exitCode=137 Oct 14 07:12:24 crc kubenswrapper[5018]: W1014 07:12:24.338082 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6719ab90_8279_40bf_9544_b35118039b0e.slice/crio-d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895 WatchSource:0}: Error finding container d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895: Status 404 returned error can't find the container with id d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.346078 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fzlkk"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.360365 5018 generic.go:334] "Generic (PLEG): container finished" podID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerID="53f8bbd9b14b8748fd6d68d48207701da31ebb8e7309185ebed668c22c0089ca" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.360642 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerDied","Data":"53f8bbd9b14b8748fd6d68d48207701da31ebb8e7309185ebed668c22c0089ca"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.394107 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.403735 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.403799 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.404212 5018 generic.go:334] "Generic (PLEG): container finished" podID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerID="f563d6fe835cc693edf98b5acc197853d2d8d196613cf1c29f3ee82afbf51623" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.404260 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf758599-jh97x" event={"ID":"dc288e96-07a9-43b9-b153-a9246dd704eb","Type":"ContainerDied","Data":"f563d6fe835cc693edf98b5acc197853d2d8d196613cf1c29f3ee82afbf51623"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.408109 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.426277 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7pr99_0f737b49-f2ba-4a7c-9b07-d934d5122363/openstack-network-exporter/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.426324 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f737b49-f2ba-4a7c-9b07-d934d5122363" containerID="3b16b9017596763bb0f70d7dd6f8aca7f1f4fbb0b4ed9d4a7005193cc2610d01" exitCode=2 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.426382 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7pr99" event={"ID":"0f737b49-f2ba-4a7c-9b07-d934d5122363","Type":"ContainerDied","Data":"3b16b9017596763bb0f70d7dd6f8aca7f1f4fbb0b4ed9d4a7005193cc2610d01"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.432778 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-169d-account-create-w72n8"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.441554 5018 generic.go:334] "Generic (PLEG): container finished" podID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerID="fb571e1ab7e3e5551c801254e6ef0537cf870849ca12640092089f0011af7c29" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.441657 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerDied","Data":"fb571e1ab7e3e5551c801254e6ef0537cf870849ca12640092089f0011af7c29"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.445064 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.169:8080/healthcheck\": read tcp 10.217.0.2:60584->10.217.0.169:8080: read: connection reset by peer" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.445192 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.169:8080/healthcheck\": read tcp 10.217.0.2:60598->10.217.0.169:8080: read: connection reset by peer" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.445267 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5hfgd"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.447381 5018 generic.go:334] "Generic (PLEG): container finished" podID="a3673da9-258f-4052-916a-90eb07abe4e6" containerID="a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.447439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerDied","Data":"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.449682 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.449701 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: E1014 07:12:24.449761 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:12:24 crc kubenswrapper[5018]: E1014 07:12:24.449794 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data podName:5ddcde8d-45fd-419b-8751-18db89bf729d nodeName:}" failed. No retries permitted until 2025-10-14 07:12:26.449781457 +0000 UTC m=+1363.033828084 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data") pod "rabbitmq-server-0" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d") : configmap "rabbitmq-config-data" not found Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.462384 5018 generic.go:334] "Generic (PLEG): container finished" podID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerID="fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd" exitCode=0 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.462444 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerDied","Data":"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.468079 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-169d-account-create-w72n8"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.468931 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1751d1e2-6a95-4a5f-b588-a58bffb94896/ovsdbserver-sb/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.468979 5018 generic.go:334] "Generic (PLEG): container finished" podID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerID="f1c52505d00c71d74b81b018f8807279db186987ee6fc9cf83ce290b691b50e9" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.469088 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerDied","Data":"f1c52505d00c71d74b81b018f8807279db186987ee6fc9cf83ce290b691b50e9"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.472733 5018 generic.go:334] "Generic (PLEG): container finished" podID="ab012259-4546-400d-9e34-217dea0ee019" containerID="d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0" exitCode=143 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.472805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerDied","Data":"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0"} Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.482600 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-55b4-account-create-72fds"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.492369 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5hfgd"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.500840 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-55b4-account-create-72fds"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.512209 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.512456 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-log" containerID="cri-o://f7f8ce51b32d6ce79cf26bf72b40b8a1faaf32d35ad244f801fb5938a444418c" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.513655 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-metadata" containerID="cri-o://6c69169eb51092731bb9a425eec4f189573257f2c5056b90bbbf411e5e89d260" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.531257 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.531478 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-log" containerID="cri-o://27b9f6c601d30ddeef6ae8d67f22e5d31f0b0856e67d3120476a411a44292642" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.531608 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-api" containerID="cri-o://620014af1412ba75636f9ca794cb656dd537c7d697efc2facc190d885dbfce9b" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: E1014 07:12:24.552257 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 07:12:24 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:12:24 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNBridge=br-int Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNEncapType=geneve Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNAvailabilityZones= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ EnableChassisAsGateway=true Oct 14 07:12:24 crc kubenswrapper[5018]: ++ PhysicalNetworks= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNHostName= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:12:24 crc kubenswrapper[5018]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:12:24 crc kubenswrapper[5018]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + cleanup_ovsdb_server_semaphore Oct 14 07:12:24 crc kubenswrapper[5018]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:12:24 crc kubenswrapper[5018]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-52nkw" message=< Oct 14 07:12:24 crc kubenswrapper[5018]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:12:24 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNBridge=br-int Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNEncapType=geneve Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNAvailabilityZones= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ EnableChassisAsGateway=true Oct 14 07:12:24 crc kubenswrapper[5018]: ++ PhysicalNetworks= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNHostName= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:12:24 crc kubenswrapper[5018]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:12:24 crc kubenswrapper[5018]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + cleanup_ovsdb_server_semaphore Oct 14 07:12:24 crc kubenswrapper[5018]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:12:24 crc kubenswrapper[5018]: > Oct 14 07:12:24 crc kubenswrapper[5018]: E1014 07:12:24.552302 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 07:12:24 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:12:24 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNBridge=br-int Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNEncapType=geneve Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNAvailabilityZones= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ EnableChassisAsGateway=true Oct 14 07:12:24 crc kubenswrapper[5018]: ++ PhysicalNetworks= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ OVNHostName= Oct 14 07:12:24 crc kubenswrapper[5018]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:12:24 crc kubenswrapper[5018]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:12:24 crc kubenswrapper[5018]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:12:24 crc kubenswrapper[5018]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + sleep 0.5 Oct 14 07:12:24 crc kubenswrapper[5018]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:12:24 crc kubenswrapper[5018]: + cleanup_ovsdb_server_semaphore Oct 14 07:12:24 crc kubenswrapper[5018]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:12:24 crc kubenswrapper[5018]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:12:24 crc kubenswrapper[5018]: > pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" containerID="cri-o://312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.552336 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" containerID="cri-o://312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" gracePeriod=29 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.564859 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.582533 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1751d1e2-6a95-4a5f-b588-a58bffb94896/ovsdbserver-sb/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.582594 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.585336 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.586180 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener-log" containerID="cri-o://3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.586278 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener" containerID="cri-o://2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.591988 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7pr99_0f737b49-f2ba-4a7c-9b07-d934d5122363/openstack-network-exporter/0.log" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.592066 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.600789 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.601046 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6d56c684b7-gtf99" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker-log" containerID="cri-o://74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.601110 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6d56c684b7-gtf99" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker" containerID="cri-o://638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.602498 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.633716 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.652786 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.652824 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.652955 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.652990 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.653011 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.653680 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.653978 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.654002 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzb74\" (UniqueName: \"kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74\") pod \"1751d1e2-6a95-4a5f-b588-a58bffb94896\" (UID: \"1751d1e2-6a95-4a5f-b588-a58bffb94896\") " Oct 14 07:12:24 crc kubenswrapper[5018]: W1014 07:12:24.654080 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60a3ff62_8fcf_47aa_b080_b94042e137be.slice/crio-b57fbb7e1a830aca6ac6460a89c08a8554b97a04fffa2dbf36dc2e534749d5db WatchSource:0}: Error finding container b57fbb7e1a830aca6ac6460a89c08a8554b97a04fffa2dbf36dc2e534749d5db: Status 404 returned error can't find the container with id b57fbb7e1a830aca6ac6460a89c08a8554b97a04fffa2dbf36dc2e534749d5db Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.655392 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts" (OuterVolumeSpecName: "scripts") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.655450 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.655691 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.663581 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.663797 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config" (OuterVolumeSpecName: "config") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.677355 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74" (OuterVolumeSpecName: "kube-api-access-pzb74") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "kube-api-access-pzb74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.682869 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="rabbitmq" containerID="cri-o://dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804" gracePeriod=604800 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.694279 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0723b8b5-790c-4dc7-874f-4e7456b8cfe8" path="/var/lib/kubelet/pods/0723b8b5-790c-4dc7-874f-4e7456b8cfe8/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.695633 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ac7a1f-c474-4f6e-9a09-e545d75e33c8" path="/var/lib/kubelet/pods/09ac7a1f-c474-4f6e-9a09-e545d75e33c8/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.696296 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f4e545d-4740-4b89-b2df-e59bf8e09144" path="/var/lib/kubelet/pods/0f4e545d-4740-4b89-b2df-e59bf8e09144/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.696769 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b2e5f4-648d-47f5-a6c3-3a6621026889" path="/var/lib/kubelet/pods/11b2e5f4-648d-47f5-a6c3-3a6621026889/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.698062 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a592b94-6c4d-4af2-99a3-860f2629332a" path="/var/lib/kubelet/pods/1a592b94-6c4d-4af2-99a3-860f2629332a/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.698833 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18" path="/var/lib/kubelet/pods/2a9d5364-4d7a-4d04-84fb-1ee0b6abdc18/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.699556 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a935d4-85ef-44c2-967b-6dbd9b7c95d8" path="/var/lib/kubelet/pods/33a935d4-85ef-44c2-967b-6dbd9b7c95d8/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.700032 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350454c7-242f-4dab-bb12-21abca9db8ce" path="/var/lib/kubelet/pods/350454c7-242f-4dab-bb12-21abca9db8ce/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.700966 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53363d98-901b-4334-bfcc-a74638255459" path="/var/lib/kubelet/pods/53363d98-901b-4334-bfcc-a74638255459/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.701437 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55477fe1-67aa-45f6-ba53-7a39c81acfd1" path="/var/lib/kubelet/pods/55477fe1-67aa-45f6-ba53-7a39c81acfd1/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.701980 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="609e45c4-103a-48d3-9113-6335839588b3" path="/var/lib/kubelet/pods/609e45c4-103a-48d3-9113-6335839588b3/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.703344 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6920d271-45c6-4d58-af8a-6707e5fa80e9" path="/var/lib/kubelet/pods/6920d271-45c6-4d58-af8a-6707e5fa80e9/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.703900 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad96861-15f8-4715-8b7c-927ae48fe59c" path="/var/lib/kubelet/pods/aad96861-15f8-4715-8b7c-927ae48fe59c/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.704694 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba453765-5fa8-4930-b9af-155c98c87d01" path="/var/lib/kubelet/pods/ba453765-5fa8-4930-b9af-155c98c87d01/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.705141 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f7d916-4a85-4465-a5fa-af904b18a50b" path="/var/lib/kubelet/pods/c4f7d916-4a85-4465-a5fa-af904b18a50b/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.706430 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7841210-eecb-475a-acf3-c1de58aa08a6" path="/var/lib/kubelet/pods/d7841210-eecb-475a-acf3-c1de58aa08a6/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.707069 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f76f5af3-9474-4186-9da0-d5ea4fe14f09" path="/var/lib/kubelet/pods/f76f5af3-9474-4186-9da0-d5ea4fe14f09/volumes" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725480 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-fab3-account-create-v5ss6"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725516 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-fab3-account-create-v5ss6"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725529 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tcn5l"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725538 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725550 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725563 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tcn5l"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725574 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725585 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725595 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvsxb"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725779 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.725978 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dfd4968f6-zjg6m" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api-log" containerID="cri-o://72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.726198 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dfd4968f6-zjg6m" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api" containerID="cri-o://f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.727474 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" containerID="cri-o://7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" gracePeriod=29 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.730075 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "c2372cde-de88-46c2-abc8-0bd1a9e7b961" (UID: "c2372cde-de88-46c2-abc8-0bd1a9e7b961"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.751926 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvsxb"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.758916 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.759163 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" containerName="nova-cell1-conductor-conductor" containerID="cri-o://ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760116 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760168 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760235 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760266 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gptwh\" (UniqueName: \"kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760300 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760329 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760366 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760462 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760476 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760501 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760524 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xphpj\" (UniqueName: \"kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj\") pod \"0f737b49-f2ba-4a7c-9b07-d934d5122363\" (UID: \"0f737b49-f2ba-4a7c-9b07-d934d5122363\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760546 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc\") pod \"dc288e96-07a9-43b9-b153-a9246dd704eb\" (UID: \"dc288e96-07a9-43b9-b153-a9246dd704eb\") " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760931 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760946 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760956 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzb74\" (UniqueName: \"kubernetes.io/projected/1751d1e2-6a95-4a5f-b588-a58bffb94896-kube-api-access-pzb74\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760965 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1751d1e2-6a95-4a5f-b588-a58bffb94896-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760974 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.760982 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2372cde-de88-46c2-abc8-0bd1a9e7b961-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.762973 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.765643 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config" (OuterVolumeSpecName: "config") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.767098 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.767297 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" containerName="nova-cell0-conductor-conductor" containerID="cri-o://3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.771339 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.771649 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="galera" containerID="cri-o://4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.784546 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.800896 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj" (OuterVolumeSpecName: "kube-api-access-xphpj") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "kube-api-access-xphpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.801008 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-skccn"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.801859 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh" (OuterVolumeSpecName: "kube-api-access-gptwh") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "kube-api-access-gptwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.808207 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.808537 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.815166 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-skccn"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.825868 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.826058 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerName="nova-scheduler-scheduler" containerID="cri-o://0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" gracePeriod=30 Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.832387 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.840009 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.859903 5018 scope.go:117] "RemoveContainer" containerID="ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.861508 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867720 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867753 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xphpj\" (UniqueName: \"kubernetes.io/projected/0f737b49-f2ba-4a7c-9b07-d934d5122363-kube-api-access-xphpj\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867762 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867771 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867780 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867790 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gptwh\" (UniqueName: \"kubernetes.io/projected/dc288e96-07a9-43b9-b153-a9246dd704eb-kube-api-access-gptwh\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867799 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867807 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f737b49-f2ba-4a7c-9b07-d934d5122363-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.867814 5018 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0f737b49-f2ba-4a7c-9b07-d934d5122363-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.868679 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.882394 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.945051 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.951900 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.969511 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.970781 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.970796 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.970805 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:24 crc kubenswrapper[5018]: I1014 07:12:24.971829 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.033057 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.034871 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.035820 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.036139 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.036198 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.040257 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.051369 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.051415 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.072348 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle\") pod \"1719e0be-9beb-4f35-b449-2a90e51ec80c\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.072447 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret\") pod \"1719e0be-9beb-4f35-b449-2a90e51ec80c\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.072542 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhw2d\" (UniqueName: \"kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d\") pod \"1719e0be-9beb-4f35-b449-2a90e51ec80c\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.072596 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config\") pod \"1719e0be-9beb-4f35-b449-2a90e51ec80c\" (UID: \"1719e0be-9beb-4f35-b449-2a90e51ec80c\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.077153 5018 scope.go:117] "RemoveContainer" containerID="55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6" Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.080027 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6\": container with ID starting with 55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6 not found: ID does not exist" containerID="55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.080079 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6"} err="failed to get container status \"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6\": rpc error: code = NotFound desc = could not find container \"55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6\": container with ID starting with 55fc5028991ab8e52ff3862f61c2e69267583970a7c60fe549504c2b4eba93a6 not found: ID does not exist" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.080107 5018 scope.go:117] "RemoveContainer" containerID="ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.080486 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "1751d1e2-6a95-4a5f-b588-a58bffb94896" (UID: "1751d1e2-6a95-4a5f-b588-a58bffb94896"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.080527 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856\": container with ID starting with ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856 not found: ID does not exist" containerID="ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.080560 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856"} err="failed to get container status \"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856\": rpc error: code = NotFound desc = could not find container \"ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856\": container with ID starting with ff21c611eec47a5f44c20a47601bbee0a99783aa663c1889f13c86cdbd99e856 not found: ID does not exist" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.101676 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.122041 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d" (OuterVolumeSpecName: "kube-api-access-mhw2d") pod "1719e0be-9beb-4f35-b449-2a90e51ec80c" (UID: "1719e0be-9beb-4f35-b449-2a90e51ec80c"). InnerVolumeSpecName "kube-api-access-mhw2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.143033 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1719e0be-9beb-4f35-b449-2a90e51ec80c" (UID: "1719e0be-9beb-4f35-b449-2a90e51ec80c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.174854 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhw2d\" (UniqueName: \"kubernetes.io/projected/1719e0be-9beb-4f35-b449-2a90e51ec80c-kube-api-access-mhw2d\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.174879 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.174888 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1751d1e2-6a95-4a5f-b588-a58bffb94896-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.180960 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.212305 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.222339 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1719e0be-9beb-4f35-b449-2a90e51ec80c" (UID: "1719e0be-9beb-4f35-b449-2a90e51ec80c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.222389 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.263257 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config" (OuterVolumeSpecName: "config") pod "dc288e96-07a9-43b9-b153-a9246dd704eb" (UID: "dc288e96-07a9-43b9-b153-a9246dd704eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.302465 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.302499 5018 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc288e96-07a9-43b9-b153-a9246dd704eb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.302511 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.323863 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "0f737b49-f2ba-4a7c-9b07-d934d5122363" (UID: "0f737b49-f2ba-4a7c-9b07-d934d5122363"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.403536 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f737b49-f2ba-4a7c-9b07-d934d5122363-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.444370 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1719e0be-9beb-4f35-b449-2a90e51ec80c" (UID: "1719e0be-9beb-4f35-b449-2a90e51ec80c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.508292 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1719e0be-9beb-4f35-b449-2a90e51ec80c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.550976 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="7464ef303afd6563d9b11c50e58be39b3a6ce6d6bc0c340a249ad132b5dfdcba" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551250 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="1c34c44d438a3d6af5f9816098723851b45124c4a7a2d7e1641ec047b86764eb" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551258 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="e58847c0dec122f46ec351ebf113ea26a71f6dd0002b59bf78203f30ea173dca" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551267 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="422e0f3e0acbeb85eabbf4a53967dd8d157d1f9364ac749656bd718ee3f5b1e7" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551331 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"7464ef303afd6563d9b11c50e58be39b3a6ce6d6bc0c340a249ad132b5dfdcba"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551364 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"1c34c44d438a3d6af5f9816098723851b45124c4a7a2d7e1641ec047b86764eb"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551377 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"e58847c0dec122f46ec351ebf113ea26a71f6dd0002b59bf78203f30ea173dca"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.551388 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"422e0f3e0acbeb85eabbf4a53967dd8d157d1f9364ac749656bd718ee3f5b1e7"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.563273 5018 generic.go:334] "Generic (PLEG): container finished" podID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerID="2b02d899d0cced29508d36139ee74410a8e6281e06ca1fb7f832e68bbfcb9048" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.563330 5018 generic.go:334] "Generic (PLEG): container finished" podID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerID="e00c8d1ad5dfcec3691b831c54fc96d156f290741de76b782f82a372f3292ae3" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.563327 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerDied","Data":"2b02d899d0cced29508d36139ee74410a8e6281e06ca1fb7f832e68bbfcb9048"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.563380 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerDied","Data":"e00c8d1ad5dfcec3691b831c54fc96d156f290741de76b782f82a372f3292ae3"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.565474 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancedba2-account-delete-fhm4g" event={"ID":"17ef95f6-1b6d-459a-b331-b114fb6205fb","Type":"ContainerStarted","Data":"be81809c6b1731e1f6afa24cb5a699dcbadb4b2ecc56482c84c86fc85ba7da58"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.568201 5018 generic.go:334] "Generic (PLEG): container finished" podID="cb738854-dc39-46ff-b965-b69efbbec865" containerID="f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.568358 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerDied","Data":"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.573345 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf758599-jh97x" event={"ID":"dc288e96-07a9-43b9-b153-a9246dd704eb","Type":"ContainerDied","Data":"1a9a8169647985e3e2103f2369d85ad8dffb90f5d9b28247131f7be1740fd5f2"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.573404 5018 scope.go:117] "RemoveContainer" containerID="f563d6fe835cc693edf98b5acc197853d2d8d196613cf1c29f3ee82afbf51623" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.573556 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf758599-jh97x" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.603858 5018 generic.go:334] "Generic (PLEG): container finished" podID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerID="27b9f6c601d30ddeef6ae8d67f22e5d31f0b0856e67d3120476a411a44292642" exitCode=143 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.603933 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerDied","Data":"27b9f6c601d30ddeef6ae8d67f22e5d31f0b0856e67d3120476a411a44292642"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.609196 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5769-account-delete-pwqc4" event={"ID":"6719ab90-8279-40bf-9544-b35118039b0e","Type":"ContainerStarted","Data":"bf2bdb19722ae792c1618512a2809aa12cbbb6af222d67173a5b54630b376efa"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.609228 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5769-account-delete-pwqc4" event={"ID":"6719ab90-8279-40bf-9544-b35118039b0e","Type":"ContainerStarted","Data":"d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895"} Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.610292 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.610682 5018 generic.go:334] "Generic (PLEG): container finished" podID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerID="74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5" exitCode=143 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.610721 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerDied","Data":"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5"} Oct 14 07:12:25 crc kubenswrapper[5018]: E1014 07:12:25.611960 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data podName:1dae2a92-e082-4271-b517-2f9219a07f2b nodeName:}" failed. No retries permitted until 2025-10-14 07:12:29.611942084 +0000 UTC m=+1366.195988711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data") pod "rabbitmq-cell1-server-0" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.647013 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.199:6080/vnc_lite.html\": dial tcp 10.217.0.199:6080: connect: connection refused" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.661785 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerDied","Data":"f7f8ce51b32d6ce79cf26bf72b40b8a1faaf32d35ad244f801fb5938a444418c"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.661960 5018 generic.go:334] "Generic (PLEG): container finished" podID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerID="f7f8ce51b32d6ce79cf26bf72b40b8a1faaf32d35ad244f801fb5938a444418c" exitCode=143 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.684708 5018 scope.go:117] "RemoveContainer" containerID="e0a7b1e3d67687b82209dd21e8cb36d14ba20c1a7b9089bb0350eb8b455c87e7" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.685048 5018 generic.go:334] "Generic (PLEG): container finished" podID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerID="72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e" exitCode=143 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.685148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerDied","Data":"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.687588 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapif4bd-account-delete-vrgxk" event={"ID":"6e0a409e-4a1f-471e-a5d5-b01213e5c0df","Type":"ContainerStarted","Data":"501eb2d4afb46bde3e9df9cadb2ab82fe475f6ae6aca2aaae27e80da253fe908"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.689507 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7pr99_0f737b49-f2ba-4a7c-9b07-d934d5122363/openstack-network-exporter/0.log" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.689556 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7pr99" event={"ID":"0f737b49-f2ba-4a7c-9b07-d934d5122363","Type":"ContainerDied","Data":"1a8e95b24b6768632df6c1d84ae1755081180d78a7737d67ae8782e519da4518"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.689717 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7pr99" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.755310 5018 generic.go:334] "Generic (PLEG): container finished" podID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" exitCode=0 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.755403 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerDied","Data":"312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.772576 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.780425 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron169d-account-delete-gjr5f" event={"ID":"60a3ff62-8fcf-47aa-b080-b94042e137be","Type":"ContainerStarted","Data":"b57fbb7e1a830aca6ac6460a89c08a8554b97a04fffa2dbf36dc2e534749d5db"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.782143 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65bf758599-jh97x"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.784496 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.795269 5018 scope.go:117] "RemoveContainer" containerID="3b16b9017596763bb0f70d7dd6f8aca7f1f4fbb0b4ed9d4a7005193cc2610d01" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.795870 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1751d1e2-6a95-4a5f-b588-a58bffb94896/ovsdbserver-sb/0.log" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.796056 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1751d1e2-6a95-4a5f-b588-a58bffb94896","Type":"ContainerDied","Data":"1e17029dd8449e4c33b88e3367a079228a62bf3acbb37159be215d24794146e3"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.796213 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.799363 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.800918 5018 generic.go:334] "Generic (PLEG): container finished" podID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerID="3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21" exitCode=143 Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.801127 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerDied","Data":"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.805388 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0fab3-account-delete-l4j7n" event={"ID":"f5d08361-7993-4757-afe3-1e9eeb5c1448","Type":"ContainerStarted","Data":"e8b1d182604f6d975d7fa3171f2d2f832dc9905428e9623298db16ffb870f3b1"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.808752 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.829966 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-7pr99"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.837213 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementd72f-account-delete-7nsjg" event={"ID":"d2c0a409-0cc4-49a4-a1ad-3de144f90e49","Type":"ContainerStarted","Data":"a2232bf814cdcc7ec222436adf4402e579a0a52be74e3aecf13426bbc1aea7b4"} Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.846556 5018 scope.go:117] "RemoveContainer" containerID="015528ec03a9ccae5d8175782d21df263addf1920d63d59b5c8dd6f6c5ecbd26" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.852952 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.857466 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.905125 5018 scope.go:117] "RemoveContainer" containerID="f1c52505d00c71d74b81b018f8807279db186987ee6fc9cf83ce290b691b50e9" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916219 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916291 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916320 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916429 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpfpk\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916553 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916588 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916638 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs\") pod \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\" (UID: \"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e\") " Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.916987 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.917936 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.933437 5018 scope.go:117] "RemoveContainer" containerID="672ba654264cd63e079885cb542e177121988e36e45b223e1019d3465e9c3dcc" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.946375 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:25 crc kubenswrapper[5018]: I1014 07:12:25.949694 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk" (OuterVolumeSpecName: "kube-api-access-lpfpk") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "kube-api-access-lpfpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.007396 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 is running failed: container process not found" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.007926 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 is running failed: container process not found" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.008485 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 is running failed: container process not found" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.008522 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" containerName="nova-cell1-conductor-conductor" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.021821 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.022035 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.022127 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.022215 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpfpk\" (UniqueName: \"kubernetes.io/projected/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-kube-api-access-lpfpk\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.285753 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data" (OuterVolumeSpecName: "config-data") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.306154 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.313847 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.318296 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" (UID: "cced7bf3-2f55-4c00-87f0-b7e17c53ee0e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.337934 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.337967 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.337978 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.337988 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.541107 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.541179 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data podName:5ddcde8d-45fd-419b-8751-18db89bf729d nodeName:}" failed. No retries permitted until 2025-10-14 07:12:30.541166009 +0000 UTC m=+1367.125212636 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data") pod "rabbitmq-server-0" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d") : configmap "rabbitmq-config-data" not found Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.595170 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 07:12:26 crc kubenswrapper[5018]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-14T07:12:24Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 14 07:12:26 crc kubenswrapper[5018]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 14 07:12:26 crc kubenswrapper[5018]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-gs786" message=< Oct 14 07:12:26 crc kubenswrapper[5018]: Exiting ovn-controller (1) [FAILED] Oct 14 07:12:26 crc kubenswrapper[5018]: Killing ovn-controller (1) [ OK ] Oct 14 07:12:26 crc kubenswrapper[5018]: 2025-10-14T07:12:24Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 14 07:12:26 crc kubenswrapper[5018]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 14 07:12:26 crc kubenswrapper[5018]: > Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.595204 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 07:12:26 crc kubenswrapper[5018]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-14T07:12:24Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 14 07:12:26 crc kubenswrapper[5018]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 14 07:12:26 crc kubenswrapper[5018]: > pod="openstack/ovn-controller-gs786" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" containerID="cri-o://76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.595242 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-gs786" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" containerID="cri-o://76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e" gracePeriod=27 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.599027 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gs786" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" probeResult="failure" output="" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.614997 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f737b49-f2ba-4a7c-9b07-d934d5122363" path="/var/lib/kubelet/pods/0f737b49-f2ba-4a7c-9b07-d934d5122363/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.615695 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14458a1a-5f03-4fff-8983-1547756eb34c" path="/var/lib/kubelet/pods/14458a1a-5f03-4fff-8983-1547756eb34c/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.616217 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1719e0be-9beb-4f35-b449-2a90e51ec80c" path="/var/lib/kubelet/pods/1719e0be-9beb-4f35-b449-2a90e51ec80c/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.617494 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" path="/var/lib/kubelet/pods/1751d1e2-6a95-4a5f-b588-a58bffb94896/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.618079 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f4e13a-f72b-46bb-92d0-32eaaf87c46b" path="/var/lib/kubelet/pods/77f4e13a-f72b-46bb-92d0-32eaaf87c46b/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.618517 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c086920-52b2-4c56-9ae6-b2c23936d783" path="/var/lib/kubelet/pods/8c086920-52b2-4c56-9ae6-b2c23936d783/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.621157 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf484b91-98a3-477a-9476-3c6f52c27297" path="/var/lib/kubelet/pods/bf484b91-98a3-477a-9476-3c6f52c27297/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.621798 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" path="/var/lib/kubelet/pods/c2372cde-de88-46c2-abc8-0bd1a9e7b961/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.622455 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" path="/var/lib/kubelet/pods/dc288e96-07a9-43b9-b153-a9246dd704eb/volumes" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.717906 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.730776 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.731054 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 is running failed: container process not found" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.731466 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 is running failed: container process not found" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.731882 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 is running failed: container process not found" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:12:26 crc kubenswrapper[5018]: E1014 07:12:26.732116 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerName="nova-scheduler-scheduler" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.748185 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.752417 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.848030 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.164:8776/healthcheck\": read tcp 10.217.0.2:51350->10.217.0.164:8776: read: connection reset by peer" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.848791 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.848958 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g597\" (UniqueName: \"kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597\") pod \"53392f65-7a1b-49fc-96a4-341f237e9288\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849012 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849137 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sflj7\" (UniqueName: \"kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849210 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849239 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849273 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs\") pod \"53392f65-7a1b-49fc-96a4-341f237e9288\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849304 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data\") pod \"467829ec-f0b0-490c-80fb-25dc08df07b4\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbmkj\" (UniqueName: \"kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj\") pod \"63753948-7b37-4eeb-a378-2947a6caa55c\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849374 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle\") pod \"467829ec-f0b0-490c-80fb-25dc08df07b4\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849411 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849450 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data\") pod \"63753948-7b37-4eeb-a378-2947a6caa55c\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849483 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849522 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs\") pod \"53392f65-7a1b-49fc-96a4-341f237e9288\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849563 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849603 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle\") pod \"53392f65-7a1b-49fc-96a4-341f237e9288\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849660 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated\") pod \"baa6d249-3441-4d16-9a74-a5232861aa21\" (UID: \"baa6d249-3441-4d16-9a74-a5232861aa21\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849696 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data\") pod \"53392f65-7a1b-49fc-96a4-341f237e9288\" (UID: \"53392f65-7a1b-49fc-96a4-341f237e9288\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849750 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbkvb\" (UniqueName: \"kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb\") pod \"467829ec-f0b0-490c-80fb-25dc08df07b4\" (UID: \"467829ec-f0b0-490c-80fb-25dc08df07b4\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.849820 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle\") pod \"63753948-7b37-4eeb-a378-2947a6caa55c\" (UID: \"63753948-7b37-4eeb-a378-2947a6caa55c\") " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.858560 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.862653 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.867823 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.869467 5018 generic.go:334] "Generic (PLEG): container finished" podID="baa6d249-3441-4d16-9a74-a5232861aa21" containerID="4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.869655 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.870773 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6mb75"] Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.870815 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerDied","Data":"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.870841 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"baa6d249-3441-4d16-9a74-a5232861aa21","Type":"ContainerDied","Data":"4c9a3e52f586edf8c75eb084c5173d36c670a666390cc158e3ece7447c634a48"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.870858 5018 scope.go:117] "RemoveContainer" containerID="4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.879355 5018 generic.go:334] "Generic (PLEG): container finished" podID="60a3ff62-8fcf-47aa-b080-b94042e137be" containerID="26faedd351dc32fc9c786f2788b0f70bcf313a89c4355df0d2a75438c2877556" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.882247 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.882778 5018 generic.go:334] "Generic (PLEG): container finished" podID="f5d08361-7993-4757-afe3-1e9eeb5c1448" containerID="710e39347beae3113a879497c3193c86b3244c3c3561dbe6d2c84ba8e00c4b07" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.883585 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597" (OuterVolumeSpecName: "kube-api-access-7g597") pod "53392f65-7a1b-49fc-96a4-341f237e9288" (UID: "53392f65-7a1b-49fc-96a4-341f237e9288"). InnerVolumeSpecName "kube-api-access-7g597". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.883754 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets" (OuterVolumeSpecName: "secrets") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.889885 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron169d-account-delete-gjr5f" event={"ID":"60a3ff62-8fcf-47aa-b080-b94042e137be","Type":"ContainerDied","Data":"26faedd351dc32fc9c786f2788b0f70bcf313a89c4355df0d2a75438c2877556"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.889924 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0fab3-account-delete-l4j7n" event={"ID":"f5d08361-7993-4757-afe3-1e9eeb5c1448","Type":"ContainerDied","Data":"710e39347beae3113a879497c3193c86b3244c3c3561dbe6d2c84ba8e00c4b07"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.899435 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7" (OuterVolumeSpecName: "kube-api-access-sflj7") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "kube-api-access-sflj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.899783 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63753948-7b37-4eeb-a378-2947a6caa55c" (UID: "63753948-7b37-4eeb-a378-2947a6caa55c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.901733 5018 generic.go:334] "Generic (PLEG): container finished" podID="17ef95f6-1b6d-459a-b331-b114fb6205fb" containerID="01983759a90d6b716e88a2a8ca9e3bdfeafcc48426b9b8276e5271cb1cc52926" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.901825 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancedba2-account-delete-fhm4g" event={"ID":"17ef95f6-1b6d-459a-b331-b114fb6205fb","Type":"ContainerDied","Data":"01983759a90d6b716e88a2a8ca9e3bdfeafcc48426b9b8276e5271cb1cc52926"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.906656 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj" (OuterVolumeSpecName: "kube-api-access-bbmkj") pod "63753948-7b37-4eeb-a378-2947a6caa55c" (UID: "63753948-7b37-4eeb-a378-2947a6caa55c"). InnerVolumeSpecName "kube-api-access-bbmkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.906680 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb" (OuterVolumeSpecName: "kube-api-access-gbkvb") pod "467829ec-f0b0-490c-80fb-25dc08df07b4" (UID: "467829ec-f0b0-490c-80fb-25dc08df07b4"). InnerVolumeSpecName "kube-api-access-gbkvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.915181 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.917815 5018 generic.go:334] "Generic (PLEG): container finished" podID="6e0a409e-4a1f-471e-a5d5-b01213e5c0df" containerID="dc592a10f1f08b665bdc5af98cb705b332dad979193cef3ba26aeda7526045bf" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.917908 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapif4bd-account-delete-vrgxk" event={"ID":"6e0a409e-4a1f-471e-a5d5-b01213e5c0df","Type":"ContainerDied","Data":"dc592a10f1f08b665bdc5af98cb705b332dad979193cef3ba26aeda7526045bf"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.926256 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6mb75"] Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.955109 5018 generic.go:334] "Generic (PLEG): container finished" podID="63753948-7b37-4eeb-a378-2947a6caa55c" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" exitCode=0 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.955150 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"63753948-7b37-4eeb-a378-2947a6caa55c","Type":"ContainerDied","Data":"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.955531 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"63753948-7b37-4eeb-a378-2947a6caa55c","Type":"ContainerDied","Data":"f1ae4098736e7c2a3598fd209ec60614ae6c1bb8f7876d466a4f1858d8ddd2ae"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.955202 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.955779 5018 scope.go:117] "RemoveContainer" containerID="3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958094 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958131 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbmkj\" (UniqueName: \"kubernetes.io/projected/63753948-7b37-4eeb-a378-2947a6caa55c-kube-api-access-bbmkj\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958141 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958168 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958177 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/baa6d249-3441-4d16-9a74-a5232861aa21-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958186 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbkvb\" (UniqueName: \"kubernetes.io/projected/467829ec-f0b0-490c-80fb-25dc08df07b4-kube-api-access-gbkvb\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958200 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958209 5018 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958219 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g597\" (UniqueName: \"kubernetes.io/projected/53392f65-7a1b-49fc-96a4-341f237e9288-kube-api-access-7g597\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958227 5018 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/baa6d249-3441-4d16-9a74-a5232861aa21-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.958236 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sflj7\" (UniqueName: \"kubernetes.io/projected/baa6d249-3441-4d16-9a74-a5232861aa21-kube-api-access-sflj7\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.967414 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.983037 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5769-account-create-487kz"] Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.989365 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gs786_ecd8d742-9217-4dd6-a035-49fab332a574/ovn-controller/0.log" Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.989405 5018 generic.go:334] "Generic (PLEG): container finished" podID="ecd8d742-9217-4dd6-a035-49fab332a574" containerID="76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e" exitCode=143 Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.989475 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786" event={"ID":"ecd8d742-9217-4dd6-a035-49fab332a574","Type":"ContainerDied","Data":"76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e"} Oct 14 07:12:26 crc kubenswrapper[5018]: I1014 07:12:26.999561 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.001231 5018 generic.go:334] "Generic (PLEG): container finished" podID="d2c0a409-0cc4-49a4-a1ad-3de144f90e49" containerID="fa1da7b22ceae7bd96055fe75e529c1ecbe80bcdf8d5305cf1c1dd8444468d8b" exitCode=0 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.001339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementd72f-account-delete-7nsjg" event={"ID":"d2c0a409-0cc4-49a4-a1ad-3de144f90e49","Type":"ContainerDied","Data":"fa1da7b22ceae7bd96055fe75e529c1ecbe80bcdf8d5305cf1c1dd8444468d8b"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.002910 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5769-account-create-487kz"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.028894 5018 generic.go:334] "Generic (PLEG): container finished" podID="6719ab90-8279-40bf-9544-b35118039b0e" containerID="bf2bdb19722ae792c1618512a2809aa12cbbb6af222d67173a5b54630b376efa" exitCode=0 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.028970 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5769-account-delete-pwqc4" event={"ID":"6719ab90-8279-40bf-9544-b35118039b0e","Type":"ContainerDied","Data":"bf2bdb19722ae792c1618512a2809aa12cbbb6af222d67173a5b54630b376efa"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.039858 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53392f65-7a1b-49fc-96a4-341f237e9288" (UID: "53392f65-7a1b-49fc-96a4-341f237e9288"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.047224 5018 generic.go:334] "Generic (PLEG): container finished" podID="53392f65-7a1b-49fc-96a4-341f237e9288" containerID="75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447" exitCode=0 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.047383 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.047377 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"53392f65-7a1b-49fc-96a4-341f237e9288","Type":"ContainerDied","Data":"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.047424 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"53392f65-7a1b-49fc-96a4-341f237e9288","Type":"ContainerDied","Data":"1cd1ec3a2c7a0f89469209f8b87ebe139e239dacdf3ec18968d24c7c7bb584f7"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.058463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" event={"ID":"cced7bf3-2f55-4c00-87f0-b7e17c53ee0e","Type":"ContainerDied","Data":"cf469bc58c9d0775ff8ae9961e65253ba16e2d367ba257c177ff52b2ba6e92d1"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.060141 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.060161 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.060518 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65c7ff6d65-cbdkc" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.061952 5018 generic.go:334] "Generic (PLEG): container finished" podID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" exitCode=0 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.061999 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"467829ec-f0b0-490c-80fb-25dc08df07b4","Type":"ContainerDied","Data":"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.062019 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"467829ec-f0b0-490c-80fb-25dc08df07b4","Type":"ContainerDied","Data":"0647c750323f18caa6a330d640f5a72168c770d7b54ec4109677bf8f679fc69c"} Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.062075 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.099772 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "467829ec-f0b0-490c-80fb-25dc08df07b4" (UID: "467829ec-f0b0-490c-80fb-25dc08df07b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.099877 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data" (OuterVolumeSpecName: "config-data") pod "467829ec-f0b0-490c-80fb-25dc08df07b4" (UID: "467829ec-f0b0-490c-80fb-25dc08df07b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.124981 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data" (OuterVolumeSpecName: "config-data") pod "53392f65-7a1b-49fc-96a4-341f237e9288" (UID: "53392f65-7a1b-49fc-96a4-341f237e9288"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.125465 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data" (OuterVolumeSpecName: "config-data") pod "63753948-7b37-4eeb-a378-2947a6caa55c" (UID: "63753948-7b37-4eeb-a378-2947a6caa55c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.125763 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "53392f65-7a1b-49fc-96a4-341f237e9288" (UID: "53392f65-7a1b-49fc-96a4-341f237e9288"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.131671 5018 scope.go:117] "RemoveContainer" containerID="4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e" Oct 14 07:12:27 crc kubenswrapper[5018]: E1014 07:12:27.132539 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e\": container with ID starting with 4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e not found: ID does not exist" containerID="4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.132571 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e"} err="failed to get container status \"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e\": rpc error: code = NotFound desc = could not find container \"4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e\": container with ID starting with 4d9b6345ec771949c259eb4654107d4b8ed92a886bffe1e867f206662216eb9e not found: ID does not exist" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.132595 5018 scope.go:117] "RemoveContainer" containerID="3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.135262 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.138906 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "baa6d249-3441-4d16-9a74-a5232861aa21" (UID: "baa6d249-3441-4d16-9a74-a5232861aa21"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: E1014 07:12:27.138921 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834\": container with ID starting with 3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834 not found: ID does not exist" containerID="3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.138978 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834"} err="failed to get container status \"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834\": rpc error: code = NotFound desc = could not find container \"3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834\": container with ID starting with 3817703118add755684d27696b62b0c1dd60ff9d3ec46e140ee92915b120d834 not found: ID does not exist" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.139005 5018 scope.go:117] "RemoveContainer" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.165923 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.165974 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467829ec-f0b0-490c-80fb-25dc08df07b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.165986 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63753948-7b37-4eeb-a378-2947a6caa55c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.165994 5018 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/baa6d249-3441-4d16-9a74-a5232861aa21-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.166003 5018 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.166011 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.166019 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.178884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "53392f65-7a1b-49fc-96a4-341f237e9288" (UID: "53392f65-7a1b-49fc-96a4-341f237e9288"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.267697 5018 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/53392f65-7a1b-49fc-96a4-341f237e9288-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.297859 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.175:9292/healthcheck\": read tcp 10.217.0.2:51548->10.217.0.175:9292: read: connection reset by peer" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.297872 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.175:9292/healthcheck\": read tcp 10.217.0.2:51546->10.217.0.175:9292: read: connection reset by peer" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.446835 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.447198 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-central-agent" containerID="cri-o://f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.447254 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="proxy-httpd" containerID="cri-o://80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.447336 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="sg-core" containerID="cri-o://d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.447388 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-notification-agent" containerID="cri-o://fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.454690 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.454937 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" containerName="kube-state-metrics" containerID="cri-o://fc84e4f07631283548a442b12f788b0fb870e48754176bcb037520224ef0cee4" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.505723 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.526550 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.536878 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.545448 5018 scope.go:117] "RemoveContainer" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" Oct 14 07:12:27 crc kubenswrapper[5018]: E1014 07:12:27.560748 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2\": container with ID starting with ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 not found: ID does not exist" containerID="ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.560803 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2"} err="failed to get container status \"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2\": rpc error: code = NotFound desc = could not find container \"ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2\": container with ID starting with ec7e3a9c262417e3affbb45408c5e01c832edd100051852aba7a17dfdc99d4f2 not found: ID does not exist" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.560833 5018 scope.go:117] "RemoveContainer" containerID="75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.572130 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.576282 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wskhd\" (UniqueName: \"kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd\") pod \"60a3ff62-8fcf-47aa-b080-b94042e137be\" (UID: \"60a3ff62-8fcf-47aa-b080-b94042e137be\") " Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.592854 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-65c7ff6d65-cbdkc"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.592983 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd" (OuterVolumeSpecName: "kube-api-access-wskhd") pod "60a3ff62-8fcf-47aa-b080-b94042e137be" (UID: "60a3ff62-8fcf-47aa-b080-b94042e137be"). InnerVolumeSpecName "kube-api-access-wskhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.611630 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.611801 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" containerName="memcached" containerID="cri-o://26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.616313 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.634675 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.655945 5018 scope.go:117] "RemoveContainer" containerID="75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447" Oct 14 07:12:27 crc kubenswrapper[5018]: E1014 07:12:27.664197 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447\": container with ID starting with 75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447 not found: ID does not exist" containerID="75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.664231 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447"} err="failed to get container status \"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447\": rpc error: code = NotFound desc = could not find container \"75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447\": container with ID starting with 75f3e1fff3f9cc6c21e4384f902ae009c6b0ff9c3543d9057ac4197522895447 not found: ID does not exist" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.664254 5018 scope.go:117] "RemoveContainer" containerID="2b02d899d0cced29508d36139ee74410a8e6281e06ca1fb7f832e68bbfcb9048" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.678821 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wskhd\" (UniqueName: \"kubernetes.io/projected/60a3ff62-8fcf-47aa-b080-b94042e137be-kube-api-access-wskhd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.705975 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.749715 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.765192 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-d7wwp"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.765222 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-hhwdt"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.775715 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-hhwdt"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.775779 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-d7wwp"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.780121 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.788041 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.793992 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.806007 5018 scope.go:117] "RemoveContainer" containerID="e00c8d1ad5dfcec3691b831c54fc96d156f290741de76b782f82a372f3292ae3" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.807992 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.808224 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5ff4499-6wpcq" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerName="keystone-api" containerID="cri-o://b483c2ab5ee4494a8c3c6374f47798895d55ff335e8b914c8d3d4714c801eb88" gracePeriod=30 Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.808794 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.824709 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-62dd-account-create-mk5wq"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.844521 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-62dd-account-create-mk5wq"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.861565 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-s4jrd"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.865694 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.879066 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-s4jrd"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.882291 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcw6w\" (UniqueName: \"kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w\") pod \"6719ab90-8279-40bf-9544-b35118039b0e\" (UID: \"6719ab90-8279-40bf-9544-b35118039b0e\") " Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.912890 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.916182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w" (OuterVolumeSpecName: "kube-api-access-rcw6w") pod "6719ab90-8279-40bf-9544-b35118039b0e" (UID: "6719ab90-8279-40bf-9544-b35118039b0e"). InnerVolumeSpecName "kube-api-access-rcw6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.926791 5018 scope.go:117] "RemoveContainer" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.934661 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-k44z4"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.947598 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f4bd-account-create-2g595"] Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.950552 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:27 crc kubenswrapper[5018]: I1014 07:12:27.985130 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-k44z4"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.005492 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btjxb\" (UniqueName: \"kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb\") pod \"17ef95f6-1b6d-459a-b331-b114fb6205fb\" (UID: \"17ef95f6-1b6d-459a-b331-b114fb6205fb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.005597 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmlbm\" (UniqueName: \"kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm\") pod \"f5d08361-7993-4757-afe3-1e9eeb5c1448\" (UID: \"f5d08361-7993-4757-afe3-1e9eeb5c1448\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.005812 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnmlf\" (UniqueName: \"kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf\") pod \"d2c0a409-0cc4-49a4-a1ad-3de144f90e49\" (UID: \"d2c0a409-0cc4-49a4-a1ad-3de144f90e49\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.006520 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcw6w\" (UniqueName: \"kubernetes.io/projected/6719ab90-8279-40bf-9544-b35118039b0e-kube-api-access-rcw6w\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.011280 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf" (OuterVolumeSpecName: "kube-api-access-jnmlf") pod "d2c0a409-0cc4-49a4-a1ad-3de144f90e49" (UID: "d2c0a409-0cc4-49a4-a1ad-3de144f90e49"). InnerVolumeSpecName "kube-api-access-jnmlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.015474 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="galera" containerID="cri-o://614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897" gracePeriod=30 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.015638 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f4bd-account-create-2g595"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.021045 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb" (OuterVolumeSpecName: "kube-api-access-btjxb") pod "17ef95f6-1b6d-459a-b331-b114fb6205fb" (UID: "17ef95f6-1b6d-459a-b331-b114fb6205fb"). InnerVolumeSpecName "kube-api-access-btjxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.034498 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm" (OuterVolumeSpecName: "kube-api-access-wmlbm") pod "f5d08361-7993-4757-afe3-1e9eeb5c1448" (UID: "f5d08361-7993-4757-afe3-1e9eeb5c1448"). InnerVolumeSpecName "kube-api-access-wmlbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.034576 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.066811 5018 scope.go:117] "RemoveContainer" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.087109 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.087812 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.108810 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmlbm\" (UniqueName: \"kubernetes.io/projected/f5d08361-7993-4757-afe3-1e9eeb5c1448-kube-api-access-wmlbm\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.108837 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnmlf\" (UniqueName: \"kubernetes.io/projected/d2c0a409-0cc4-49a4-a1ad-3de144f90e49-kube-api-access-jnmlf\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.108851 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btjxb\" (UniqueName: \"kubernetes.io/projected/17ef95f6-1b6d-459a-b331-b114fb6205fb-kube-api-access-btjxb\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.117146 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983\": container with ID starting with 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 not found: ID does not exist" containerID="0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.117190 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983"} err="failed to get container status \"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983\": rpc error: code = NotFound desc = could not find container \"0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983\": container with ID starting with 0b9d77d0a4212c0a665d7f9f3efe4e9fd32a8c1840549f2922226e0aa3735983 not found: ID does not exist" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.208131 5018 generic.go:334] "Generic (PLEG): container finished" podID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerID="84742c9fbf39092ccf977df6e5a65995900f3bd0bbaf07ff5f7aed100b72b67f" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.208218 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerDied","Data":"84742c9fbf39092ccf977df6e5a65995900f3bd0bbaf07ff5f7aed100b72b67f"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213168 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213241 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213294 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqrsh\" (UniqueName: \"kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213369 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213398 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213421 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213466 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213484 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213526 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213543 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213581 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213602 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pkhw\" (UniqueName: \"kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213676 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs\") pod \"ab012259-4546-400d-9e34-217dea0ee019\" (UID: \"ab012259-4546-400d-9e34-217dea0ee019\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213711 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213726 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.213747 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts\") pod \"a3673da9-258f-4052-916a-90eb07abe4e6\" (UID: \"a3673da9-258f-4052-916a-90eb07abe4e6\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.214736 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs" (OuterVolumeSpecName: "logs") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.214953 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs" (OuterVolumeSpecName: "logs") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.217906 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw" (OuterVolumeSpecName: "kube-api-access-6pkhw") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "kube-api-access-6pkhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.227206 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.240013 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts" (OuterVolumeSpecName: "scripts") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.247753 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-dfd4968f6-zjg6m" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:57628->10.217.0.157:9311: read: connection reset by peer" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.247764 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-dfd4968f6-zjg6m" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:57626->10.217.0.157:9311: read: connection reset by peer" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.248015 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh" (OuterVolumeSpecName: "kube-api-access-tqrsh") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "kube-api-access-tqrsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.248298 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.251465 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts" (OuterVolumeSpecName: "scripts") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.254044 5018 generic.go:334] "Generic (PLEG): container finished" podID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerID="6c69169eb51092731bb9a425eec4f189573257f2c5056b90bbbf411e5e89d260" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.254108 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerDied","Data":"6c69169eb51092731bb9a425eec4f189573257f2c5056b90bbbf411e5e89d260"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.269028 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementd72f-account-delete-7nsjg" event={"ID":"d2c0a409-0cc4-49a4-a1ad-3de144f90e49","Type":"ContainerDied","Data":"a2232bf814cdcc7ec222436adf4402e579a0a52be74e3aecf13426bbc1aea7b4"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.269087 5018 scope.go:117] "RemoveContainer" containerID="fa1da7b22ceae7bd96055fe75e529c1ecbe80bcdf8d5305cf1c1dd8444468d8b" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.269205 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementd72f-account-delete-7nsjg" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.271431 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.272797 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancedba2-account-delete-fhm4g" event={"ID":"17ef95f6-1b6d-459a-b331-b114fb6205fb","Type":"ContainerDied","Data":"be81809c6b1731e1f6afa24cb5a699dcbadb4b2ecc56482c84c86fc85ba7da58"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.272913 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancedba2-account-delete-fhm4g" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.273546 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.275586 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.275628 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.286683 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0fab3-account-delete-l4j7n" event={"ID":"f5d08361-7993-4757-afe3-1e9eeb5c1448","Type":"ContainerDied","Data":"e8b1d182604f6d975d7fa3171f2d2f832dc9905428e9623298db16ffb870f3b1"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.286774 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0fab3-account-delete-l4j7n" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.307164 5018 generic.go:334] "Generic (PLEG): container finished" podID="ab012259-4546-400d-9e34-217dea0ee019" containerID="6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.307211 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7649bf8b6b-5szh6" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.307246 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerDied","Data":"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.307269 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7649bf8b6b-5szh6" event={"ID":"ab012259-4546-400d-9e34-217dea0ee019","Type":"ContainerDied","Data":"fcc6bfc97a2b39a7c251e57a72e8f5d749619f8452b5d6864d9ec92f5c3efbd8"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.311372 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5769-account-delete-pwqc4" event={"ID":"6719ab90-8279-40bf-9544-b35118039b0e","Type":"ContainerDied","Data":"d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.311405 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2dd6b66714671f2353a16b192b99630ad106794369bfe85143667092d516895" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.311475 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5769-account-delete-pwqc4" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315829 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3673da9-258f-4052-916a-90eb07abe4e6-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315854 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3673da9-258f-4052-916a-90eb07abe4e6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315862 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315870 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab012259-4546-400d-9e34-217dea0ee019-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315880 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqrsh\" (UniqueName: \"kubernetes.io/projected/a3673da9-258f-4052-916a-90eb07abe4e6-kube-api-access-tqrsh\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315888 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315895 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.315903 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pkhw\" (UniqueName: \"kubernetes.io/projected/ab012259-4546-400d-9e34-217dea0ee019-kube-api-access-6pkhw\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.323931 5018 generic.go:334] "Generic (PLEG): container finished" podID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerID="b96379567c4eaad728194e40c42a49fa57be113eff3ccca7f60a2b9c2df42bbb" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.323966 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerDied","Data":"b96379567c4eaad728194e40c42a49fa57be113eff3ccca7f60a2b9c2df42bbb"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.327343 5018 generic.go:334] "Generic (PLEG): container finished" podID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerID="80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.327369 5018 generic.go:334] "Generic (PLEG): container finished" podID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerID="d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" exitCode=2 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.327401 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerDied","Data":"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.327421 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerDied","Data":"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.334761 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron169d-account-delete-gjr5f" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.334766 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron169d-account-delete-gjr5f" event={"ID":"60a3ff62-8fcf-47aa-b080-b94042e137be","Type":"ContainerDied","Data":"b57fbb7e1a830aca6ac6460a89c08a8554b97a04fffa2dbf36dc2e534749d5db"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.340117 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.344263 5018 generic.go:334] "Generic (PLEG): container finished" podID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" containerID="fc84e4f07631283548a442b12f788b0fb870e48754176bcb037520224ef0cee4" exitCode=2 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.344332 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe","Type":"ContainerDied","Data":"fc84e4f07631283548a442b12f788b0fb870e48754176bcb037520224ef0cee4"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.346538 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data" (OuterVolumeSpecName: "config-data") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.360407 5018 generic.go:334] "Generic (PLEG): container finished" podID="a3673da9-258f-4052-916a-90eb07abe4e6" containerID="b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.360464 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerDied","Data":"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.360491 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a3673da9-258f-4052-916a-90eb07abe4e6","Type":"ContainerDied","Data":"94af7b16c95a3ae7f24f31c266c8993c1adb8a361512fbf5e0be6ee10506b46d"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.360548 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.362689 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.367210 5018 generic.go:334] "Generic (PLEG): container finished" podID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerID="620014af1412ba75636f9ca794cb656dd537c7d697efc2facc190d885dbfce9b" exitCode=0 Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.367271 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerDied","Data":"620014af1412ba75636f9ca794cb656dd537c7d697efc2facc190d885dbfce9b"} Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.390888 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data" (OuterVolumeSpecName: "config-data") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.426738 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.426761 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.426773 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.426781 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.467322 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a3673da9-258f-4052-916a-90eb07abe4e6" (UID: "a3673da9-258f-4052-916a-90eb07abe4e6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.518756 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.528067 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.528097 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3673da9-258f-4052-916a-90eb07abe4e6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.539437 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b076446_7046_4b7a_b315_f8a560d5604b.slice/crio-conmon-3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75b77f89_1956_43e0_b6df_dd4d390e4cef.slice/crio-conmon-f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.565684 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.568631 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab012259-4546-400d-9e34-217dea0ee019" (UID: "ab012259-4546-400d-9e34-217dea0ee019"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.596320 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 is running failed: container process not found" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.596731 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 is running failed: container process not found" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.597060 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 is running failed: container process not found" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.597088 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" containerName="nova-cell0-conductor-conductor" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.613629 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-75695fb6c7-7ztlf" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.165:9696/\": dial tcp 10.217.0.165:9696: connect: connection refused" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.617854 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1801677b-8aa6-4c0f-9f88-66bea7da4835" path="/var/lib/kubelet/pods/1801677b-8aa6-4c0f-9f88-66bea7da4835/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.619041 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" path="/var/lib/kubelet/pods/467829ec-f0b0-490c-80fb-25dc08df07b4/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.619576 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c45b3a9-3966-4584-ac73-138eed247f9d" path="/var/lib/kubelet/pods/4c45b3a9-3966-4584-ac73-138eed247f9d/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.620078 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" path="/var/lib/kubelet/pods/53392f65-7a1b-49fc-96a4-341f237e9288/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.621332 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e80324-584c-4cfd-b71e-11adfe900797" path="/var/lib/kubelet/pods/57e80324-584c-4cfd-b71e-11adfe900797/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.622198 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" path="/var/lib/kubelet/pods/63753948-7b37-4eeb-a378-2947a6caa55c/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.622695 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652c0172-b19f-4bfa-9f82-3ad458cbc8cd" path="/var/lib/kubelet/pods/652c0172-b19f-4bfa-9f82-3ad458cbc8cd/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.623676 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22" path="/var/lib/kubelet/pods/8f2aa868-3d6a-4ff9-8f40-bb12bf6e4a22/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.624112 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b736e7b0-e21d-4859-8c8f-033c5a62e7c8" path="/var/lib/kubelet/pods/b736e7b0-e21d-4859-8c8f-033c5a62e7c8/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.624781 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" path="/var/lib/kubelet/pods/baa6d249-3441-4d16-9a74-a5232861aa21/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.625813 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" path="/var/lib/kubelet/pods/cced7bf3-2f55-4c00-87f0-b7e17c53ee0e/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.626484 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e474ca2f-3868-45b7-9251-4e254c8c14dd" path="/var/lib/kubelet/pods/e474ca2f-3868-45b7-9251-4e254c8c14dd/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.626979 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdbc6001-fbcd-4162-b1ab-633b49b15ae9" path="/var/lib/kubelet/pods/fdbc6001-fbcd-4162-b1ab-633b49b15ae9/volumes" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.637017 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.638783 5018 scope.go:117] "RemoveContainer" containerID="01983759a90d6b716e88a2a8ca9e3bdfeafcc48426b9b8276e5271cb1cc52926" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.639207 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab012259-4546-400d-9e34-217dea0ee019-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.669747 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gs786_ecd8d742-9217-4dd6-a035-49fab332a574/ovn-controller/0.log" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.669815 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.703659 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.714727 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.736358 5018 scope.go:117] "RemoveContainer" containerID="710e39347beae3113a879497c3193c86b3244c3c3561dbe6d2c84ba8e00c4b07" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.736787 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.743478 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.747837 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5frrj\" (UniqueName: \"kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.747880 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.748143 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.748218 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.748241 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.748272 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.748298 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts\") pod \"ecd8d742-9217-4dd6-a035-49fab332a574\" (UID: \"ecd8d742-9217-4dd6-a035-49fab332a574\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.749229 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.752043 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.752109 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run" (OuterVolumeSpecName: "var-run") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.753465 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts" (OuterVolumeSpecName: "scripts") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.758536 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj" (OuterVolumeSpecName: "kube-api-access-5frrj") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "kube-api-access-5frrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.758761 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.788800 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.803364 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancedba2-account-delete-fhm4g"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.808642 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.809238 5018 scope.go:117] "RemoveContainer" containerID="6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.814771 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementd72f-account-delete-7nsjg"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.850843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.850888 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.850908 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs\") pod \"c6a69b66-5d0d-410b-9d2e-6179c8079077\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.850942 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.850970 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851000 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rqwl\" (UniqueName: \"kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl\") pod \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851024 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle\") pod \"c6a69b66-5d0d-410b-9d2e-6179c8079077\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851052 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851081 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851108 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs5x6\" (UniqueName: \"kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.851128 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.853117 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.856544 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts" (OuterVolumeSpecName: "scripts") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.861958 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs" (OuterVolumeSpecName: "logs") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862310 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config\") pod \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862337 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pbm4\" (UniqueName: \"kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862378 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f66l\" (UniqueName: \"kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l\") pod \"c6a69b66-5d0d-410b-9d2e-6179c8079077\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97g5w\" (UniqueName: \"kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862433 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862464 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862950 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle\") pod \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.862971 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs\") pod \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\" (UID: \"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863021 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs\") pod \"c6a69b66-5d0d-410b-9d2e-6179c8079077\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863045 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863063 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863107 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863123 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data\") pod \"c6a69b66-5d0d-410b-9d2e-6179c8079077\" (UID: \"c6a69b66-5d0d-410b-9d2e-6179c8079077\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863141 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863254 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863275 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863320 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs\") pod \"248e38fb-736b-4333-a184-5a6fdd04daeb\" (UID: \"248e38fb-736b-4333-a184-5a6fdd04daeb\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863347 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863369 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs\") pod \"52261d3b-8ca7-4632-9f05-a8e5b656f397\" (UID: \"52261d3b-8ca7-4632-9f05-a8e5b656f397\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.863386 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"dc86ed1d-ae09-4120-8e15-2a960334a853\" (UID: \"dc86ed1d-ae09-4120-8e15-2a960334a853\") " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864386 5018 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864403 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864411 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864420 5018 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864428 5018 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd8d742-9217-4dd6-a035-49fab332a574-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864436 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd8d742-9217-4dd6-a035-49fab332a574-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864445 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248e38fb-736b-4333-a184-5a6fdd04daeb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.864455 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5frrj\" (UniqueName: \"kubernetes.io/projected/ecd8d742-9217-4dd6-a035-49fab332a574-kube-api-access-5frrj\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.876280 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl" (OuterVolumeSpecName: "kube-api-access-5rqwl") pod "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" (UID: "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe"). InnerVolumeSpecName "kube-api-access-5rqwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.879045 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.884446 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs" (OuterVolumeSpecName: "logs") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.891601 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l" (OuterVolumeSpecName: "kube-api-access-4f66l") pod "c6a69b66-5d0d-410b-9d2e-6179c8079077" (UID: "c6a69b66-5d0d-410b-9d2e-6179c8079077"). InnerVolumeSpecName "kube-api-access-4f66l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.891696 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.896613 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.900763 5018 scope.go:117] "RemoveContainer" containerID="d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.909403 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0fab3-account-delete-l4j7n"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.909454 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.910781 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs" (OuterVolumeSpecName: "logs") pod "c6a69b66-5d0d-410b-9d2e-6179c8079077" (UID: "c6a69b66-5d0d-410b-9d2e-6179c8079077"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.911103 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs" (OuterVolumeSpecName: "logs") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.919756 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7649bf8b6b-5szh6"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.921147 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w" (OuterVolumeSpecName: "kube-api-access-97g5w") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "kube-api-access-97g5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.924901 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.926557 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts" (OuterVolumeSpecName: "scripts") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.926855 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.926952 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4" (OuterVolumeSpecName: "kube-api-access-7pbm4") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "kube-api-access-7pbm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.927341 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.928836 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecd8d742-9217-4dd6-a035-49fab332a574" (UID: "ecd8d742-9217-4dd6-a035-49fab332a574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.929844 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" (UID: "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.930536 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6" (OuterVolumeSpecName: "kube-api-access-bs5x6") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "kube-api-access-bs5x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.930751 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.931643 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data" (OuterVolumeSpecName: "config-data") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.933778 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6a69b66-5d0d-410b-9d2e-6179c8079077" (UID: "c6a69b66-5d0d-410b-9d2e-6179c8079077"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.936106 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder5769-account-delete-pwqc4"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.943699 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.945698 5018 scope.go:117] "RemoveContainer" containerID="6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.952941 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.953799 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f\": container with ID starting with 6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f not found: ID does not exist" containerID="6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.953860 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f"} err="failed to get container status \"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f\": rpc error: code = NotFound desc = could not find container \"6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f\": container with ID starting with 6eba9b07e8d470039fc68bfdc176e5c278974381633c9d65e7b70bbf1debff0f not found: ID does not exist" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.953894 5018 scope.go:117] "RemoveContainer" containerID="d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0" Oct 14 07:12:28 crc kubenswrapper[5018]: E1014 07:12:28.955688 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0\": container with ID starting with d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0 not found: ID does not exist" containerID="d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.955732 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0"} err="failed to get container status \"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0\": rpc error: code = NotFound desc = could not find container \"d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0\": container with ID starting with d79286b556bbc3217e24a3bad8fcf0fef1f0a00672049b6ab78e5ccc9ab81bd0 not found: ID does not exist" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.955757 5018 scope.go:117] "RemoveContainer" containerID="26faedd351dc32fc9c786f2788b0f70bcf313a89c4355df0d2a75438c2877556" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.960678 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron169d-account-delete-gjr5f"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.966757 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967877 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a69b66-5d0d-410b-9d2e-6179c8079077-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967896 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967903 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967913 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc86ed1d-ae09-4120-8e15-2a960334a853-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967928 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967948 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967956 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52261d3b-8ca7-4632-9f05-a8e5b656f397-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967968 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967977 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967985 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.967993 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rqwl\" (UniqueName: \"kubernetes.io/projected/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-api-access-5rqwl\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968001 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968009 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968017 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs5x6\" (UniqueName: \"kubernetes.io/projected/52261d3b-8ca7-4632-9f05-a8e5b656f397-kube-api-access-bs5x6\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968026 5018 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968036 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pbm4\" (UniqueName: \"kubernetes.io/projected/248e38fb-736b-4333-a184-5a6fdd04daeb-kube-api-access-7pbm4\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968047 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecd8d742-9217-4dd6-a035-49fab332a574-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968056 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f66l\" (UniqueName: \"kubernetes.io/projected/c6a69b66-5d0d-410b-9d2e-6179c8079077-kube-api-access-4f66l\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.968065 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97g5w\" (UniqueName: \"kubernetes.io/projected/dc86ed1d-ae09-4120-8e15-2a960334a853-kube-api-access-97g5w\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.970781 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data" (OuterVolumeSpecName: "config-data") pod "c6a69b66-5d0d-410b-9d2e-6179c8079077" (UID: "c6a69b66-5d0d-410b-9d2e-6179c8079077"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.971744 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:12:28 crc kubenswrapper[5018]: I1014 07:12:28.978776 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.031832 5018 scope.go:117] "RemoveContainer" containerID="b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.069437 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w58jp\" (UniqueName: \"kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.069749 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.069861 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.069979 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.070154 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.070298 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.070389 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs\") pod \"75b77f89-1956-43e0-b6df-dd4d390e4cef\" (UID: \"75b77f89-1956-43e0-b6df-dd4d390e4cef\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.071310 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.073038 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.074889 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.075937 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs" (OuterVolumeSpecName: "logs") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.079367 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.081431 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" (UID: "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.085208 5018 scope.go:117] "RemoveContainer" containerID="a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.099787 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.118841 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp" (OuterVolumeSpecName: "kube-api-access-w58jp") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "kube-api-access-w58jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.158113 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.162093 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.172792 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.172828 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.172873 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.173000 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.173052 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.173580 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gjp2\" (UniqueName: \"kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.173653 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.173676 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle\") pod \"8d9387da-03d5-46e4-b442-aecb5fd1e766\" (UID: \"8d9387da-03d5-46e4-b442-aecb5fd1e766\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174071 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w58jp\" (UniqueName: \"kubernetes.io/projected/75b77f89-1956-43e0-b6df-dd4d390e4cef-kube-api-access-w58jp\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174090 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b77f89-1956-43e0-b6df-dd4d390e4cef-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174100 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174114 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174123 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174132 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174139 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174148 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174728 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.174951 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.175919 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" (UID: "66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.182848 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2" (OuterVolumeSpecName: "kube-api-access-7gjp2") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "kube-api-access-7gjp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.186930 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts" (OuterVolumeSpecName: "scripts") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.218259 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.219319 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "248e38fb-736b-4333-a184-5a6fdd04daeb" (UID: "248e38fb-736b-4333-a184-5a6fdd04daeb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.235062 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.256727 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data" (OuterVolumeSpecName: "config-data") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.257290 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.259162 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.265093 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c6a69b66-5d0d-410b-9d2e-6179c8079077" (UID: "c6a69b66-5d0d-410b-9d2e-6179c8079077"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.273108 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data" (OuterVolumeSpecName: "config-data") pod "52261d3b-8ca7-4632-9f05-a8e5b656f397" (UID: "52261d3b-8ca7-4632-9f05-a8e5b656f397"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275104 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/248e38fb-736b-4333-a184-5a6fdd04daeb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275119 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275128 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gjp2\" (UniqueName: \"kubernetes.io/projected/8d9387da-03d5-46e4-b442-aecb5fd1e766-kube-api-access-7gjp2\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275138 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275148 5018 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275158 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275168 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52261d3b-8ca7-4632-9f05-a8e5b656f397-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275176 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275184 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275192 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a69b66-5d0d-410b-9d2e-6179c8079077-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275200 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275208 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.275216 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9387da-03d5-46e4-b442-aecb5fd1e766-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.284774 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data" (OuterVolumeSpecName: "config-data") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.318208 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "75b77f89-1956-43e0-b6df-dd4d390e4cef" (UID: "75b77f89-1956-43e0-b6df-dd4d390e4cef"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.323121 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.341529 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dc86ed1d-ae09-4120-8e15-2a960334a853" (UID: "dc86ed1d-ae09-4120-8e15-2a960334a853"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.358184 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.377528 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.377564 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.377579 5018 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.377591 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b77f89-1956-43e0-b6df-dd4d390e4cef-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.377602 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc86ed1d-ae09-4120-8e15-2a960334a853-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.395690 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data" (OuterVolumeSpecName: "config-data") pod "8d9387da-03d5-46e4-b442-aecb5fd1e766" (UID: "8d9387da-03d5-46e4-b442-aecb5fd1e766"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.453119 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.453593 5018 scope.go:117] "RemoveContainer" containerID="b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.454789 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6\": container with ID starting with b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6 not found: ID does not exist" containerID="b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.454814 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6"} err="failed to get container status \"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6\": rpc error: code = NotFound desc = could not find container \"b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6\": container with ID starting with b37b52ae7a016d59054dd0b90ee49ab6f0cdecd65d2ff5340c5a804ce4283ba6 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.454839 5018 scope.go:117] "RemoveContainer" containerID="a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.455995 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09\": container with ID starting with a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09 not found: ID does not exist" containerID="a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.456017 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09"} err="failed to get container status \"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09\": rpc error: code = NotFound desc = could not find container \"a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09\": container with ID starting with a7cafca19c7a0b9e566194749e9f87738599977ecebce7915df8573544dbab09 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.457112 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.463778 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.478511 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9387da-03d5-46e4-b442-aecb5fd1e766-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.505294 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.505367 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6a69b66-5d0d-410b-9d2e-6179c8079077","Type":"ContainerDied","Data":"6b777a2685bad24921e479b5e2015642fb370e2734cce4591099a6578f181128"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.505453 5018 scope.go:117] "RemoveContainer" containerID="6c69169eb51092731bb9a425eec4f189573257f2c5056b90bbbf411e5e89d260" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.514241 5018 generic.go:334] "Generic (PLEG): container finished" podID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerID="f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6" exitCode=0 Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.514292 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dfd4968f6-zjg6m" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.514304 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerDied","Data":"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.514674 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dfd4968f6-zjg6m" event={"ID":"75b77f89-1956-43e0-b6df-dd4d390e4cef","Type":"ContainerDied","Data":"0ae60f1d8e72562e05a13c1170f1faa3e21e2b6e67f2b67664643d1b931f0f7a"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.517493 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gs786_ecd8d742-9217-4dd6-a035-49fab332a574/ovn-controller/0.log" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.517679 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gs786" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.517788 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gs786" event={"ID":"ecd8d742-9217-4dd6-a035-49fab332a574","Type":"ContainerDied","Data":"6b35061bda013c035c4d6f3e6ae981f891709a14cd9e86badf29fade130b68b1"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.521730 5018 generic.go:334] "Generic (PLEG): container finished" podID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" containerID="26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c" exitCode=0 Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.521796 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e","Type":"ContainerDied","Data":"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.521815 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e","Type":"ContainerDied","Data":"80330bc7a1cf897e49fc60598b544559c0f8ffe3226288e1802790d454d95900"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.521883 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.530644 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe","Type":"ContainerDied","Data":"fdc65b82ae3f4227b60de78688b45336f41c5f90388edef59fe4b3514d8cc591"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.530720 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539038 5018 generic.go:334] "Generic (PLEG): container finished" podID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerID="fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" exitCode=0 Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539062 5018 generic.go:334] "Generic (PLEG): container finished" podID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerID="f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" exitCode=0 Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539089 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539116 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerDied","Data":"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539160 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerDied","Data":"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.539170 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9387da-03d5-46e4-b442-aecb5fd1e766","Type":"ContainerDied","Data":"2e29f4391ae9949b53ecb8add8a020c15b469a585747b89932d114a30311cbb0"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.543436 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248e38fb-736b-4333-a184-5a6fdd04daeb","Type":"ContainerDied","Data":"5bebb8f7a6fb0607c3475c0663ea8e8f358eea574d7a1f95b900d716594ff4aa"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.543483 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.552716 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.553406 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"dc86ed1d-ae09-4120-8e15-2a960334a853","Type":"ContainerDied","Data":"20b83ce71f8b4a06e3e250e42d4eee3fab91a87919112570e5e5a88ce4c5b716"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.559037 5018 generic.go:334] "Generic (PLEG): container finished" podID="5b076446-7046-4b7a-b315-f8a560d5604b" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" exitCode=0 Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.559098 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5b076446-7046-4b7a-b315-f8a560d5604b","Type":"ContainerDied","Data":"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.559125 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5b076446-7046-4b7a-b315-f8a560d5604b","Type":"ContainerDied","Data":"f06f3e9d0065ec29bdf951365c7634fb8922e9d418f7018bd4afa39e89b5df38"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.559181 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.560275 5018 scope.go:117] "RemoveContainer" containerID="f7f8ce51b32d6ce79cf26bf72b40b8a1faaf32d35ad244f801fb5938a444418c" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.562989 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52261d3b-8ca7-4632-9f05-a8e5b656f397","Type":"ContainerDied","Data":"536abeac8032c0588c7d065bccc2a3860b6b60af148dd561a4d43bf1181ed82f"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.563075 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.566821 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapif4bd-account-delete-vrgxk" event={"ID":"6e0a409e-4a1f-471e-a5d5-b01213e5c0df","Type":"ContainerDied","Data":"501eb2d4afb46bde3e9df9cadb2ab82fe475f6ae6aca2aaae27e80da253fe908"} Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.566847 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501eb2d4afb46bde3e9df9cadb2ab82fe475f6ae6aca2aaae27e80da253fe908" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.566898 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapif4bd-account-delete-vrgxk" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581119 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs\") pod \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581260 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data\") pod \"5b076446-7046-4b7a-b315-f8a560d5604b\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581294 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq\") pod \"6e0a409e-4a1f-471e-a5d5-b01213e5c0df\" (UID: \"6e0a409e-4a1f-471e-a5d5-b01213e5c0df\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581442 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data\") pod \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581478 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config\") pod \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581633 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle\") pod \"5b076446-7046-4b7a-b315-f8a560d5604b\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581678 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq8dc\" (UniqueName: \"kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc\") pod \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581720 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle\") pod \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\" (UID: \"c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.581766 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jth68\" (UniqueName: \"kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68\") pod \"5b076446-7046-4b7a-b315-f8a560d5604b\" (UID: \"5b076446-7046-4b7a-b315-f8a560d5604b\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.582590 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data" (OuterVolumeSpecName: "config-data") pod "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" (UID: "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.592392 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" (UID: "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.593013 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68" (OuterVolumeSpecName: "kube-api-access-jth68") pod "5b076446-7046-4b7a-b315-f8a560d5604b" (UID: "5b076446-7046-4b7a-b315-f8a560d5604b"). InnerVolumeSpecName "kube-api-access-jth68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.600247 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq" (OuterVolumeSpecName: "kube-api-access-nk4wq") pod "6e0a409e-4a1f-471e-a5d5-b01213e5c0df" (UID: "6e0a409e-4a1f-471e-a5d5-b01213e5c0df"). InnerVolumeSpecName "kube-api-access-nk4wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.609101 5018 scope.go:117] "RemoveContainer" containerID="f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.613656 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc" (OuterVolumeSpecName: "kube-api-access-vq8dc") pod "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" (UID: "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e"). InnerVolumeSpecName "kube-api-access-vq8dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.635064 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b076446-7046-4b7a-b315-f8a560d5604b" (UID: "5b076446-7046-4b7a-b315-f8a560d5604b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.636867 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" (UID: "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.637427 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" (UID: "c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.638751 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.646394 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.649758 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data" (OuterVolumeSpecName: "config-data") pod "5b076446-7046-4b7a-b315-f8a560d5604b" (UID: "5b076446-7046-4b7a-b315-f8a560d5604b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.654914 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.660191 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.673160 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.683652 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.683719 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data podName:1dae2a92-e082-4271-b517-2f9219a07f2b nodeName:}" failed. No retries permitted until 2025-10-14 07:12:37.683701961 +0000 UTC m=+1374.267748648 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data") pod "rabbitmq-cell1-server-0" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684713 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk4wq\" (UniqueName: \"kubernetes.io/projected/6e0a409e-4a1f-471e-a5d5-b01213e5c0df-kube-api-access-nk4wq\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684743 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684777 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684786 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684796 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq8dc\" (UniqueName: \"kubernetes.io/projected/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-kube-api-access-vq8dc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684805 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684815 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jth68\" (UniqueName: \"kubernetes.io/projected/5b076446-7046-4b7a-b315-f8a560d5604b-kube-api-access-jth68\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684823 5018 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.684844 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b076446-7046-4b7a-b315-f8a560d5604b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.685757 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.695245 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.701750 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-dfd4968f6-zjg6m"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.710148 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.711353 5018 scope.go:117] "RemoveContainer" containerID="72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.724534 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.731669 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.740259 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.746760 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.750887 5018 scope.go:117] "RemoveContainer" containerID="f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.752386 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6\": container with ID starting with f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6 not found: ID does not exist" containerID="f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.752424 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6"} err="failed to get container status \"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6\": rpc error: code = NotFound desc = could not find container \"f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6\": container with ID starting with f6ba9f81fff3a0df599460603664c4a48cf828694301943d0c6fdd12afc504c6 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.752448 5018 scope.go:117] "RemoveContainer" containerID="72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.752897 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gs786"] Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.754588 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e\": container with ID starting with 72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e not found: ID does not exist" containerID="72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.754639 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e"} err="failed to get container status \"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e\": rpc error: code = NotFound desc = could not find container \"72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e\": container with ID starting with 72f599bf2a8f9e057d1e8206f5fc2e126b74c9c4fff65c90da9ed3c17e2a554e not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.754662 5018 scope.go:117] "RemoveContainer" containerID="76c82c4a12d6a47a3590016ce8ec76fbc803f50385766ce1d769751b6f61039e" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.759184 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.765728 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.785472 5018 scope.go:117] "RemoveContainer" containerID="26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.815365 5018 scope.go:117] "RemoveContainer" containerID="26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.815838 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c\": container with ID starting with 26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c not found: ID does not exist" containerID="26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.815874 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c"} err="failed to get container status \"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c\": rpc error: code = NotFound desc = could not find container \"26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c\": container with ID starting with 26cb43ea58227c7a4af9d242a4d9898a54abc5f129843b9d1e9e55327eb6210c not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.815899 5018 scope.go:117] "RemoveContainer" containerID="fc84e4f07631283548a442b12f788b0fb870e48754176bcb037520224ef0cee4" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.832972 5018 scope.go:117] "RemoveContainer" containerID="80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.851845 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.857809 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.860056 5018 scope.go:117] "RemoveContainer" containerID="d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.875694 5018 scope.go:117] "RemoveContainer" containerID="fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.891946 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.896468 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.911294 5018 scope.go:117] "RemoveContainer" containerID="f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.915839 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.920606 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapif4bd-account-delete-vrgxk"] Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.922339 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c736e3c0-1c92-463c-aa8e-104d6bb2c6c0/ovn-northd/0.log" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.922394 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.931775 5018 scope.go:117] "RemoveContainer" containerID="80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.932122 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2\": container with ID starting with 80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2 not found: ID does not exist" containerID="80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932190 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2"} err="failed to get container status \"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2\": rpc error: code = NotFound desc = could not find container \"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2\": container with ID starting with 80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932223 5018 scope.go:117] "RemoveContainer" containerID="d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.932488 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6\": container with ID starting with d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6 not found: ID does not exist" containerID="d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932519 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6"} err="failed to get container status \"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6\": rpc error: code = NotFound desc = could not find container \"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6\": container with ID starting with d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932537 5018 scope.go:117] "RemoveContainer" containerID="fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.932860 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29\": container with ID starting with fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29 not found: ID does not exist" containerID="fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932891 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29"} err="failed to get container status \"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29\": rpc error: code = NotFound desc = could not find container \"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29\": container with ID starting with fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.932909 5018 scope.go:117] "RemoveContainer" containerID="f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" Oct 14 07:12:29 crc kubenswrapper[5018]: E1014 07:12:29.933107 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698\": container with ID starting with f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698 not found: ID does not exist" containerID="f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933134 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698"} err="failed to get container status \"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698\": rpc error: code = NotFound desc = could not find container \"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698\": container with ID starting with f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933153 5018 scope.go:117] "RemoveContainer" containerID="80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933522 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2"} err="failed to get container status \"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2\": rpc error: code = NotFound desc = could not find container \"80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2\": container with ID starting with 80607b63698c4abfd890edfe51254df34896e898a31df8ed1ad52beaa03523f2 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933548 5018 scope.go:117] "RemoveContainer" containerID="d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933882 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6"} err="failed to get container status \"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6\": rpc error: code = NotFound desc = could not find container \"d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6\": container with ID starting with d58d797cbc268a4d99d915fe8462cfeaa3456a271af4c21e835efcd94bd64fc6 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.933910 5018 scope.go:117] "RemoveContainer" containerID="fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.934158 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29"} err="failed to get container status \"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29\": rpc error: code = NotFound desc = could not find container \"fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29\": container with ID starting with fea7ae0b766bb946b3acc17906438c26ea6e2299559ff862a7dae98932c89c29 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.934183 5018 scope.go:117] "RemoveContainer" containerID="f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.934462 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698"} err="failed to get container status \"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698\": rpc error: code = NotFound desc = could not find container \"f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698\": container with ID starting with f3a41418f366f3f684a869a03ce86a21cb5657e3380337f3a1b086480f14c698 not found: ID does not exist" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.934491 5018 scope.go:117] "RemoveContainer" containerID="620014af1412ba75636f9ca794cb656dd537c7d697efc2facc190d885dbfce9b" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.979596 5018 scope.go:117] "RemoveContainer" containerID="27b9f6c601d30ddeef6ae8d67f22e5d31f0b0856e67d3120476a411a44292642" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989463 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr476\" (UniqueName: \"kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989557 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989723 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989743 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989820 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.989836 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle\") pod \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\" (UID: \"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0\") " Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.990361 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.992694 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config" (OuterVolumeSpecName: "config") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:29 crc kubenswrapper[5018]: I1014 07:12:29.992897 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts" (OuterVolumeSpecName: "scripts") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.000564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476" (OuterVolumeSpecName: "kube-api-access-vr476") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "kube-api-access-vr476". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.000996 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.001530 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.001926 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.001960 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.008259 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.010148 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.015583 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.017999 5018 scope.go:117] "RemoveContainer" containerID="84742c9fbf39092ccf977df6e5a65995900f3bd0bbaf07ff5f7aed100b72b67f" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.018130 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.018167 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.038996 5018 scope.go:117] "RemoveContainer" containerID="fb571e1ab7e3e5551c801254e6ef0537cf870849ca12640092089f0011af7c29" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.067915 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.072601 5018 scope.go:117] "RemoveContainer" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.080771 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" (UID: "c736e3c0-1c92-463c-aa8e-104d6bb2c6c0"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091719 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091740 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091748 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091757 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091765 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091773 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.091782 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr476\" (UniqueName: \"kubernetes.io/projected/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0-kube-api-access-vr476\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.310832 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.336665 5018 scope.go:117] "RemoveContainer" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.337051 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178\": container with ID starting with 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 not found: ID does not exist" containerID="3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.337087 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178"} err="failed to get container status \"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178\": rpc error: code = NotFound desc = could not find container \"3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178\": container with ID starting with 3b1d6afa2cf8b4c7e36018ec7bfec996d61bee6c68cca576c1ee4c35e40c1178 not found: ID does not exist" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.337115 5018 scope.go:117] "RemoveContainer" containerID="b96379567c4eaad728194e40c42a49fa57be113eff3ccca7f60a2b9c2df42bbb" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.348977 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.366028 5018 scope.go:117] "RemoveContainer" containerID="53f8bbd9b14b8748fd6d68d48207701da31ebb8e7309185ebed668c22c0089ca" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.473069 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.583978 5018 generic.go:334] "Generic (PLEG): container finished" podID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerID="db5a58f0d517ed36e8022372c1d0cbb0ac63dbc1e9c0b3d9627b4e9590b35c60" exitCode=0 Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.584052 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerDied","Data":"db5a58f0d517ed36e8022372c1d0cbb0ac63dbc1e9c0b3d9627b4e9590b35c60"} Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601328 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerID="614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897" exitCode=0 Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601392 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerDied","Data":"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897"} Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6ca7068c-e306-467d-972d-6ec8439d2b85","Type":"ContainerDied","Data":"a6d658ec166fe9a0201d6b6a684b09e0b7e1e06c96418a5221797ee370089267"} Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601437 5018 scope.go:117] "RemoveContainer" containerID="614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601434 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601493 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601518 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601550 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601566 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601571 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnpkb\" (UniqueName: \"kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601613 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601689 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601725 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.601869 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config\") pod \"6ca7068c-e306-467d-972d-6ec8439d2b85\" (UID: \"6ca7068c-e306-467d-972d-6ec8439d2b85\") " Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.602253 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.602299 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data podName:5ddcde8d-45fd-419b-8751-18db89bf729d nodeName:}" failed. No retries permitted until 2025-10-14 07:12:38.602285404 +0000 UTC m=+1375.186332031 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data") pod "rabbitmq-server-0" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d") : configmap "rabbitmq-config-data" not found Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.602678 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.605329 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.606489 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.606520 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.608285 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb" (OuterVolumeSpecName: "kube-api-access-tnpkb") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "kube-api-access-tnpkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.608857 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets" (OuterVolumeSpecName: "secrets") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.617492 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "mysql-db") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.626035 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c736e3c0-1c92-463c-aa8e-104d6bb2c6c0/ovn-northd/0.log" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.626080 5018 generic.go:334] "Generic (PLEG): container finished" podID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" exitCode=139 Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.626183 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.630340 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ef95f6-1b6d-459a-b331-b114fb6205fb" path="/var/lib/kubelet/pods/17ef95f6-1b6d-459a-b331-b114fb6205fb/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.631937 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" path="/var/lib/kubelet/pods/248e38fb-736b-4333-a184-5a6fdd04daeb/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.633853 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.635532 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" path="/var/lib/kubelet/pods/52261d3b-8ca7-4632-9f05-a8e5b656f397/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.636267 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" path="/var/lib/kubelet/pods/5b076446-7046-4b7a-b315-f8a560d5604b/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.638383 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.638444 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a3ff62-8fcf-47aa-b080-b94042e137be" path="/var/lib/kubelet/pods/60a3ff62-8fcf-47aa-b080-b94042e137be/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.639018 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" path="/var/lib/kubelet/pods/66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.639446 5018 scope.go:117] "RemoveContainer" containerID="92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.639648 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6719ab90-8279-40bf-9544-b35118039b0e" path="/var/lib/kubelet/pods/6719ab90-8279-40bf-9544-b35118039b0e/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.640140 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0a409e-4a1f-471e-a5d5-b01213e5c0df" path="/var/lib/kubelet/pods/6e0a409e-4a1f-471e-a5d5-b01213e5c0df/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.641151 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" path="/var/lib/kubelet/pods/75b77f89-1956-43e0-b6df-dd4d390e4cef/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.652193 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" path="/var/lib/kubelet/pods/8d9387da-03d5-46e4-b442-aecb5fd1e766/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.652997 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" path="/var/lib/kubelet/pods/a3673da9-258f-4052-916a-90eb07abe4e6/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.654791 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab012259-4546-400d-9e34-217dea0ee019" path="/var/lib/kubelet/pods/ab012259-4546-400d-9e34-217dea0ee019/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.669294 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" path="/var/lib/kubelet/pods/c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.670034 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" path="/var/lib/kubelet/pods/c6a69b66-5d0d-410b-9d2e-6179c8079077/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.670824 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2c0a409-0cc4-49a4-a1ad-3de144f90e49" path="/var/lib/kubelet/pods/d2c0a409-0cc4-49a4-a1ad-3de144f90e49/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.671493 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "6ca7068c-e306-467d-972d-6ec8439d2b85" (UID: "6ca7068c-e306-467d-972d-6ec8439d2b85"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.671832 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" path="/var/lib/kubelet/pods/dc86ed1d-ae09-4120-8e15-2a960334a853/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.672559 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" path="/var/lib/kubelet/pods/ecd8d742-9217-4dd6-a035-49fab332a574/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.673041 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d08361-7993-4757-afe3-1e9eeb5c1448" path="/var/lib/kubelet/pods/f5d08361-7993-4757-afe3-1e9eeb5c1448/volumes" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.673783 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerDied","Data":"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321"} Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.673813 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c736e3c0-1c92-463c-aa8e-104d6bb2c6c0","Type":"ContainerDied","Data":"d121869711c7268cef9749d0de08bc8fc5607bbc0e7fd583dd9ac0b2cb48a58b"} Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.682889 5018 scope.go:117] "RemoveContainer" containerID="614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.683525 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897\": container with ID starting with 614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897 not found: ID does not exist" containerID="614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.683573 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897"} err="failed to get container status \"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897\": rpc error: code = NotFound desc = could not find container \"614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897\": container with ID starting with 614f72c401b14dd23a871d07173f2d14da7deb3b069807220f72ed7386e13897 not found: ID does not exist" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.684414 5018 scope.go:117] "RemoveContainer" containerID="92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.684702 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be\": container with ID starting with 92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be not found: ID does not exist" containerID="92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.684733 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be"} err="failed to get container status \"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be\": rpc error: code = NotFound desc = could not find container \"92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be\": container with ID starting with 92db2cbf31a286c538a3e75dc468b3264a30c044c5c8bb6a552029e826f1a8be not found: ID does not exist" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.684754 5018 scope.go:117] "RemoveContainer" containerID="9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.687140 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.692537 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703130 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703216 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703243 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703286 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703341 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703373 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703407 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8f69\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703465 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703511 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703556 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.703595 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf\") pod \"1dae2a92-e082-4271-b517-2f9219a07f2b\" (UID: \"1dae2a92-e082-4271-b517-2f9219a07f2b\") " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704048 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704068 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6ca7068c-e306-467d-972d-6ec8439d2b85-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704083 5018 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704095 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704176 5018 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704214 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca7068c-e306-467d-972d-6ec8439d2b85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704235 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704315 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704331 5018 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ca7068c-e306-467d-972d-6ec8439d2b85-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.704342 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnpkb\" (UniqueName: \"kubernetes.io/projected/6ca7068c-e306-467d-972d-6ec8439d2b85-kube-api-access-tnpkb\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.707729 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.707988 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69" (OuterVolumeSpecName: "kube-api-access-q8f69") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "kube-api-access-q8f69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.708048 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.712653 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.712778 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.712869 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info" (OuterVolumeSpecName: "pod-info") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.714791 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.720263 5018 scope.go:117] "RemoveContainer" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.728201 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.740437 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data" (OuterVolumeSpecName: "config-data") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.741604 5018 scope.go:117] "RemoveContainer" containerID="9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.741976 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a\": container with ID starting with 9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a not found: ID does not exist" containerID="9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.742013 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a"} err="failed to get container status \"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a\": rpc error: code = NotFound desc = could not find container \"9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a\": container with ID starting with 9749926f58ed06ee1f1742e08400b9fbe78a768127ce31a02ece865ea9bb4b7a not found: ID does not exist" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.742037 5018 scope.go:117] "RemoveContainer" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" Oct 14 07:12:30 crc kubenswrapper[5018]: E1014 07:12:30.742283 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321\": container with ID starting with 93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321 not found: ID does not exist" containerID="93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.742310 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321"} err="failed to get container status \"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321\": rpc error: code = NotFound desc = could not find container \"93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321\": container with ID starting with 93dad57ea08cdace4897d3af1b026fbfa17cb1387d6e96d1352ce20819fac321 not found: ID does not exist" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.753715 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf" (OuterVolumeSpecName: "server-conf") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.798522 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1dae2a92-e082-4271-b517-2f9219a07f2b" (UID: "1dae2a92-e082-4271-b517-2f9219a07f2b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805203 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805236 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805246 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805255 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1dae2a92-e082-4271-b517-2f9219a07f2b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805263 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1dae2a92-e082-4271-b517-2f9219a07f2b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805273 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805281 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805291 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1dae2a92-e082-4271-b517-2f9219a07f2b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805300 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805307 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805328 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.805337 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8f69\" (UniqueName: \"kubernetes.io/projected/1dae2a92-e082-4271-b517-2f9219a07f2b-kube-api-access-q8f69\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.819751 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.906335 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.932489 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:12:30 crc kubenswrapper[5018]: I1014 07:12:30.940321 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.267001 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/keystone-5ff4499-6wpcq" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerName="keystone-api" probeResult="failure" output="Get \"https://10.217.0.150:5000/v3\": read tcp 10.217.0.2:44692->10.217.0.150:5000: read: connection reset by peer" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.426000 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.517390 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.517438 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.517887 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.517958 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518297 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518343 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518375 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518415 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518541 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518600 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518647 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.518678 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9xx\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx\") pod \"5ddcde8d-45fd-419b-8751-18db89bf729d\" (UID: \"5ddcde8d-45fd-419b-8751-18db89bf729d\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.519047 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.522903 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.523460 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.523908 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx" (OuterVolumeSpecName: "kube-api-access-7q9xx") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "kube-api-access-7q9xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.529458 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.537737 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.568724 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info" (OuterVolumeSpecName: "pod-info") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.569114 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.574345 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data" (OuterVolumeSpecName: "config-data") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.590130 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf" (OuterVolumeSpecName: "server-conf") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620564 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5ddcde8d-45fd-419b-8751-18db89bf729d-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620598 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620636 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620649 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ddcde8d-45fd-419b-8751-18db89bf729d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620658 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620667 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620678 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620686 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5ddcde8d-45fd-419b-8751-18db89bf729d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.620694 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9xx\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-kube-api-access-7q9xx\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.638298 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.647131 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5ddcde8d-45fd-419b-8751-18db89bf729d" (UID: "5ddcde8d-45fd-419b-8751-18db89bf729d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.652098 5018 generic.go:334] "Generic (PLEG): container finished" podID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerID="b483c2ab5ee4494a8c3c6374f47798895d55ff335e8b914c8d3d4714c801eb88" exitCode=0 Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.652192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ff4499-6wpcq" event={"ID":"390ed4c4-a917-47a0-83e5-5c05fbaf2005","Type":"ContainerDied","Data":"b483c2ab5ee4494a8c3c6374f47798895d55ff335e8b914c8d3d4714c801eb88"} Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.656370 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1dae2a92-e082-4271-b517-2f9219a07f2b","Type":"ContainerDied","Data":"d63779d331926f3a753bff10e4496ef456d30c4baa486cde1f2edcf14ed6d9ae"} Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.656428 5018 scope.go:117] "RemoveContainer" containerID="db5a58f0d517ed36e8022372c1d0cbb0ac63dbc1e9c0b3d9627b4e9590b35c60" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.656612 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.662391 5018 generic.go:334] "Generic (PLEG): container finished" podID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerID="dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804" exitCode=0 Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.662432 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerDied","Data":"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804"} Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.662473 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5ddcde8d-45fd-419b-8751-18db89bf729d","Type":"ContainerDied","Data":"f348120b2e00e1e900a787ef0d2a2e30b118ad5602a3c48534dbce41a0172cc7"} Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.662640 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.688836 5018 scope.go:117] "RemoveContainer" containerID="124f8aee2f1028a966a6895f7aa8cc6cea30f99e63669865c80d68ed595f9d7f" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.689213 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.714523 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.724650 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5ddcde8d-45fd-419b-8751-18db89bf729d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.724689 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.726454 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.726502 5018 scope.go:117] "RemoveContainer" containerID="dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.731326 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.739598 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.759821 5018 scope.go:117] "RemoveContainer" containerID="6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.809494 5018 scope.go:117] "RemoveContainer" containerID="dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804" Oct 14 07:12:31 crc kubenswrapper[5018]: E1014 07:12:31.813737 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804\": container with ID starting with dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804 not found: ID does not exist" containerID="dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.813777 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804"} err="failed to get container status \"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804\": rpc error: code = NotFound desc = could not find container \"dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804\": container with ID starting with dc4f4e9fd4df270595f23cdfeadfe6030624a7f69468c8853a21ea544e79f804 not found: ID does not exist" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.813802 5018 scope.go:117] "RemoveContainer" containerID="6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a" Oct 14 07:12:31 crc kubenswrapper[5018]: E1014 07:12:31.817694 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a\": container with ID starting with 6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a not found: ID does not exist" containerID="6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.817719 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a"} err="failed to get container status \"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a\": rpc error: code = NotFound desc = could not find container \"6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a\": container with ID starting with 6916fb0a68b9430aa663605c6f2a575359bf12cb3d91350095ec038016c8597a not found: ID does not exist" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826032 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826117 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826145 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml26k\" (UniqueName: \"kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826173 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826210 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826234 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826248 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.826264 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys\") pod \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\" (UID: \"390ed4c4-a917-47a0-83e5-5c05fbaf2005\") " Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.837739 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.843132 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts" (OuterVolumeSpecName: "scripts") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.843161 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.859799 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k" (OuterVolumeSpecName: "kube-api-access-ml26k") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "kube-api-access-ml26k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.887805 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data" (OuterVolumeSpecName: "config-data") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.894752 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.897814 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927437 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927470 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927479 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927488 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927498 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml26k\" (UniqueName: \"kubernetes.io/projected/390ed4c4-a917-47a0-83e5-5c05fbaf2005-kube-api-access-ml26k\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927508 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.927516 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:31 crc kubenswrapper[5018]: I1014 07:12:31.929708 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "390ed4c4-a917-47a0-83e5-5c05fbaf2005" (UID: "390ed4c4-a917-47a0-83e5-5c05fbaf2005"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.028487 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/390ed4c4-a917-47a0-83e5-5c05fbaf2005-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.463401 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.463726 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.463767 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.464818 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.464874 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d" gracePeriod=600 Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.579719 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.604511 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.614975 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" path="/var/lib/kubelet/pods/1dae2a92-e082-4271-b517-2f9219a07f2b/volumes" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.616008 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" path="/var/lib/kubelet/pods/5ddcde8d-45fd-419b-8751-18db89bf729d/volumes" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.617118 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" path="/var/lib/kubelet/pods/6ca7068c-e306-467d-972d-6ec8439d2b85/volumes" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.617861 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" path="/var/lib/kubelet/pods/c736e3c0-1c92-463c-aa8e-104d6bb2c6c0/volumes" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647189 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs\") pod \"a49d558a-c4f6-49d3-8210-c38799ab89c4\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647245 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs\") pod \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647300 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data\") pod \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647335 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle\") pod \"a49d558a-c4f6-49d3-8210-c38799ab89c4\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647364 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data\") pod \"a49d558a-c4f6-49d3-8210-c38799ab89c4\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647393 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle\") pod \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647427 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqsbx\" (UniqueName: \"kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx\") pod \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom\") pod \"a49d558a-c4f6-49d3-8210-c38799ab89c4\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647487 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom\") pod \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\" (UID: \"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.647505 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhk5w\" (UniqueName: \"kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w\") pod \"a49d558a-c4f6-49d3-8210-c38799ab89c4\" (UID: \"a49d558a-c4f6-49d3-8210-c38799ab89c4\") " Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.648280 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs" (OuterVolumeSpecName: "logs") pod "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" (UID: "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.648292 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs" (OuterVolumeSpecName: "logs") pod "a49d558a-c4f6-49d3-8210-c38799ab89c4" (UID: "a49d558a-c4f6-49d3-8210-c38799ab89c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.656387 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w" (OuterVolumeSpecName: "kube-api-access-dhk5w") pod "a49d558a-c4f6-49d3-8210-c38799ab89c4" (UID: "a49d558a-c4f6-49d3-8210-c38799ab89c4"). InnerVolumeSpecName "kube-api-access-dhk5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.659011 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" (UID: "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.666465 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a49d558a-c4f6-49d3-8210-c38799ab89c4" (UID: "a49d558a-c4f6-49d3-8210-c38799ab89c4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.673002 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx" (OuterVolumeSpecName: "kube-api-access-zqsbx") pod "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" (UID: "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a"). InnerVolumeSpecName "kube-api-access-zqsbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.693751 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a49d558a-c4f6-49d3-8210-c38799ab89c4" (UID: "a49d558a-c4f6-49d3-8210-c38799ab89c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.714957 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" (UID: "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.716796 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d" exitCode=0 Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.717112 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.717159 5018 scope.go:117] "RemoveContainer" containerID="cdea188a48901d02c9f343bc0ed72b7ee7245d6bd39dae8562ad03103b7c4ecd" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.721487 5018 generic.go:334] "Generic (PLEG): container finished" podID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerID="638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45" exitCode=0 Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.721672 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d56c684b7-gtf99" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.721612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerDied","Data":"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.722063 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d56c684b7-gtf99" event={"ID":"a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a","Type":"ContainerDied","Data":"20b6b92a2174a2ef0fe0c4f5b0cf89dc09edc6f8f9a30dd835bb36436c9d3209"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.723137 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ff4499-6wpcq" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.723155 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ff4499-6wpcq" event={"ID":"390ed4c4-a917-47a0-83e5-5c05fbaf2005","Type":"ContainerDied","Data":"5aa41398fc676eceb74ac034937b5650afc9d535a7d583a4d5d68e3dd17ce981"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.723684 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data" (OuterVolumeSpecName: "config-data") pod "a49d558a-c4f6-49d3-8210-c38799ab89c4" (UID: "a49d558a-c4f6-49d3-8210-c38799ab89c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.728218 5018 generic.go:334] "Generic (PLEG): container finished" podID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerID="2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7" exitCode=0 Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.728257 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerDied","Data":"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.728289 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" event={"ID":"a49d558a-c4f6-49d3-8210-c38799ab89c4","Type":"ContainerDied","Data":"e00074b1a0eed139d4d4c3237a8c52c563bac8cdbc8c7530d7301cf8141f8925"} Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.731920 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d9d59b97b-ng574" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.746094 5018 scope.go:117] "RemoveContainer" containerID="638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751284 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751320 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751357 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhk5w\" (UniqueName: \"kubernetes.io/projected/a49d558a-c4f6-49d3-8210-c38799ab89c4-kube-api-access-dhk5w\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751370 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a49d558a-c4f6-49d3-8210-c38799ab89c4-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751383 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751394 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751405 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a49d558a-c4f6-49d3-8210-c38799ab89c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751416 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751427 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqsbx\" (UniqueName: \"kubernetes.io/projected/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-kube-api-access-zqsbx\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.751487 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.759172 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5ff4499-6wpcq"] Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.778253 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data" (OuterVolumeSpecName: "config-data") pod "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" (UID: "a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.780950 5018 scope.go:117] "RemoveContainer" containerID="74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.788528 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.792433 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5d9d59b97b-ng574"] Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.810902 5018 scope.go:117] "RemoveContainer" containerID="638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45" Oct 14 07:12:32 crc kubenswrapper[5018]: E1014 07:12:32.813004 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45\": container with ID starting with 638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45 not found: ID does not exist" containerID="638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.813049 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45"} err="failed to get container status \"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45\": rpc error: code = NotFound desc = could not find container \"638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45\": container with ID starting with 638f022a499802c8c4705b89780ef56dbadf632acdb01723951fadcb4802bd45 not found: ID does not exist" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.813078 5018 scope.go:117] "RemoveContainer" containerID="74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5" Oct 14 07:12:32 crc kubenswrapper[5018]: E1014 07:12:32.813358 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5\": container with ID starting with 74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5 not found: ID does not exist" containerID="74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.813382 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5"} err="failed to get container status \"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5\": rpc error: code = NotFound desc = could not find container \"74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5\": container with ID starting with 74798ee03c22dba6573a275a0898f8d8c187eefd961c4bc99989ed2eafd0f6c5 not found: ID does not exist" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.813396 5018 scope.go:117] "RemoveContainer" containerID="b483c2ab5ee4494a8c3c6374f47798895d55ff335e8b914c8d3d4714c801eb88" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.834955 5018 scope.go:117] "RemoveContainer" containerID="2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.852519 5018 scope.go:117] "RemoveContainer" containerID="3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.852730 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.873669 5018 scope.go:117] "RemoveContainer" containerID="2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7" Oct 14 07:12:32 crc kubenswrapper[5018]: E1014 07:12:32.873997 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7\": container with ID starting with 2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7 not found: ID does not exist" containerID="2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.874024 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7"} err="failed to get container status \"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7\": rpc error: code = NotFound desc = could not find container \"2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7\": container with ID starting with 2c9e93a0240d999026b4cbe8526120d8049ed39603a79f1747a8f075270db3e7 not found: ID does not exist" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.874043 5018 scope.go:117] "RemoveContainer" containerID="3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21" Oct 14 07:12:32 crc kubenswrapper[5018]: E1014 07:12:32.874380 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21\": container with ID starting with 3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21 not found: ID does not exist" containerID="3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21" Oct 14 07:12:32 crc kubenswrapper[5018]: I1014 07:12:32.874399 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21"} err="failed to get container status \"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21\": rpc error: code = NotFound desc = could not find container \"3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21\": container with ID starting with 3ea07b5eb737c599c905cd7176fb51ed2e093576a223c198b71fc8ae54c56a21 not found: ID does not exist" Oct 14 07:12:33 crc kubenswrapper[5018]: I1014 07:12:33.058665 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:12:33 crc kubenswrapper[5018]: I1014 07:12:33.066578 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6d56c684b7-gtf99"] Oct 14 07:12:33 crc kubenswrapper[5018]: I1014 07:12:33.745926 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb"} Oct 14 07:12:34 crc kubenswrapper[5018]: I1014 07:12:34.623997 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" path="/var/lib/kubelet/pods/390ed4c4-a917-47a0-83e5-5c05fbaf2005/volumes" Oct 14 07:12:34 crc kubenswrapper[5018]: I1014 07:12:34.626666 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" path="/var/lib/kubelet/pods/a49d558a-c4f6-49d3-8210-c38799ab89c4/volumes" Oct 14 07:12:34 crc kubenswrapper[5018]: I1014 07:12:34.630849 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" path="/var/lib/kubelet/pods/a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a/volumes" Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.000825 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.001329 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.001664 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.001691 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.005353 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.008866 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.011267 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:35 crc kubenswrapper[5018]: E1014 07:12:35.011347 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.000852 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.002361 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.002834 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.002930 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.002998 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.005240 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.006851 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.006914 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.579430 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594364 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594405 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr9pc\" (UniqueName: \"kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594452 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594474 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594498 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594517 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.594559 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs\") pod \"08aa3b2a-b639-4e29-9593-0da33b053fbd\" (UID: \"08aa3b2a-b639-4e29-9593-0da33b053fbd\") " Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.619988 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc" (OuterVolumeSpecName: "kube-api-access-jr9pc") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "kube-api-access-jr9pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.620068 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.654798 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.675099 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.681914 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.689924 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config" (OuterVolumeSpecName: "config") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.692362 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "08aa3b2a-b639-4e29-9593-0da33b053fbd" (UID: "08aa3b2a-b639-4e29-9593-0da33b053fbd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696331 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr9pc\" (UniqueName: \"kubernetes.io/projected/08aa3b2a-b639-4e29-9593-0da33b053fbd-kube-api-access-jr9pc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696438 5018 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696521 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696552 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696609 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696658 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.696676 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08aa3b2a-b639-4e29-9593-0da33b053fbd-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.848257 5018 generic.go:334] "Generic (PLEG): container finished" podID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerID="d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a" exitCode=0 Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.848494 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerDied","Data":"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a"} Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.848673 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695fb6c7-7ztlf" event={"ID":"08aa3b2a-b639-4e29-9593-0da33b053fbd","Type":"ContainerDied","Data":"5a1f482ad48c64cd1008be1b1ffec437fe4b450cfba9ff6cf964744b364053e6"} Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.848708 5018 scope.go:117] "RemoveContainer" containerID="fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.848600 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695fb6c7-7ztlf" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.876541 5018 scope.go:117] "RemoveContainer" containerID="d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.890153 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.894160 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-75695fb6c7-7ztlf"] Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.907341 5018 scope.go:117] "RemoveContainer" containerID="fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd" Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.907930 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd\": container with ID starting with fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd not found: ID does not exist" containerID="fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.907990 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd"} err="failed to get container status \"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd\": rpc error: code = NotFound desc = could not find container \"fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd\": container with ID starting with fe5fa5a4c5e0b2b31b3a8a03cf306b955dacd5f7d66582645fba3c01bee433cd not found: ID does not exist" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.908014 5018 scope.go:117] "RemoveContainer" containerID="d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a" Oct 14 07:12:40 crc kubenswrapper[5018]: E1014 07:12:40.908316 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a\": container with ID starting with d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a not found: ID does not exist" containerID="d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a" Oct 14 07:12:40 crc kubenswrapper[5018]: I1014 07:12:40.908333 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a"} err="failed to get container status \"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a\": rpc error: code = NotFound desc = could not find container \"d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a\": container with ID starting with d5fa9ec9293551b72d8ff687bf42afe2dc1e255a90fb2c42337251649622a73a not found: ID does not exist" Oct 14 07:12:42 crc kubenswrapper[5018]: I1014 07:12:42.621207 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" path="/var/lib/kubelet/pods/08aa3b2a-b639-4e29-9593-0da33b053fbd/volumes" Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.001207 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.002326 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.003110 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.003178 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.003839 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.007369 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.009884 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:45 crc kubenswrapper[5018]: E1014 07:12:45.009985 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.000903 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.001687 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.002101 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.002146 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.005877 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.011056 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.014058 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:12:50 crc kubenswrapper[5018]: E1014 07:12:50.014112 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-52nkw" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.006005 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024225 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024268 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xgj8\" (UniqueName: \"kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024295 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024342 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024359 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024380 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom\") pod \"cb738854-dc39-46ff-b965-b69efbbec865\" (UID: \"cb738854-dc39-46ff-b965-b69efbbec865\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.024695 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.027499 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-52nkw_17e34896-a3ff-4810-a21e-b8c5ccad8379/ovs-vswitchd/0.log" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.028330 5018 generic.go:334] "Generic (PLEG): container finished" podID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" exitCode=137 Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.028422 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerDied","Data":"7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4"} Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.041924 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts" (OuterVolumeSpecName: "scripts") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.043981 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.048010 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8" (OuterVolumeSpecName: "kube-api-access-2xgj8") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "kube-api-access-2xgj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.059767 5018 generic.go:334] "Generic (PLEG): container finished" podID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerID="f75caf8dbbd7c4743e32fa90f2af02023f5f583c5e27a58a4d2ae6216d5c09b9" exitCode=137 Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.059957 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"f75caf8dbbd7c4743e32fa90f2af02023f5f583c5e27a58a4d2ae6216d5c09b9"} Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.062583 5018 generic.go:334] "Generic (PLEG): container finished" podID="cb738854-dc39-46ff-b965-b69efbbec865" containerID="73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc" exitCode=137 Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.062638 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerDied","Data":"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc"} Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.062687 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb738854-dc39-46ff-b965-b69efbbec865","Type":"ContainerDied","Data":"8e44d9b1f49487838ec1b4b24cc4fa83c15f2cce33a69ca620f79e26adfac29e"} Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.062704 5018 scope.go:117] "RemoveContainer" containerID="f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.063088 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.078431 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.081667 5018 scope.go:117] "RemoveContainer" containerID="73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.105255 5018 scope.go:117] "RemoveContainer" containerID="f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f" Oct 14 07:12:54 crc kubenswrapper[5018]: E1014 07:12:54.105765 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f\": container with ID starting with f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f not found: ID does not exist" containerID="f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.105811 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f"} err="failed to get container status \"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f\": rpc error: code = NotFound desc = could not find container \"f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f\": container with ID starting with f4b3e94408372fab5655fb3b77c1945481ec45216d7d8906b0a52391ec00e25f not found: ID does not exist" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.105840 5018 scope.go:117] "RemoveContainer" containerID="73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc" Oct 14 07:12:54 crc kubenswrapper[5018]: E1014 07:12:54.106315 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc\": container with ID starting with 73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc not found: ID does not exist" containerID="73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.106344 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc"} err="failed to get container status \"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc\": rpc error: code = NotFound desc = could not find container \"73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc\": container with ID starting with 73518b6e36a5f24f231ff4479726d7b080baccec9e22789498e2a12dd720b0dc not found: ID does not exist" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.125758 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.125797 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xgj8\" (UniqueName: \"kubernetes.io/projected/cb738854-dc39-46ff-b965-b69efbbec865-kube-api-access-2xgj8\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.125811 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.125820 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb738854-dc39-46ff-b965-b69efbbec865-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.125829 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.154752 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data" (OuterVolumeSpecName: "config-data") pod "cb738854-dc39-46ff-b965-b69efbbec865" (UID: "cb738854-dc39-46ff-b965-b69efbbec865"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.226614 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb738854-dc39-46ff-b965-b69efbbec865-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.305375 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-52nkw_17e34896-a3ff-4810-a21e-b8c5ccad8379/ovs-vswitchd/0.log" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.306410 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328002 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9gd9\" (UniqueName: \"kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328118 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328151 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328198 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328224 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run" (OuterVolumeSpecName: "var-run") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328250 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328278 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328341 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts\") pod \"17e34896-a3ff-4810-a21e-b8c5ccad8379\" (UID: \"17e34896-a3ff-4810-a21e-b8c5ccad8379\") " Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328313 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log" (OuterVolumeSpecName: "var-log") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328381 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib" (OuterVolumeSpecName: "var-lib") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328717 5018 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328730 5018 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328740 5018 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-log\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.328747 5018 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17e34896-a3ff-4810-a21e-b8c5ccad8379-var-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.329692 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts" (OuterVolumeSpecName: "scripts") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.333091 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9" (OuterVolumeSpecName: "kube-api-access-p9gd9") pod "17e34896-a3ff-4810-a21e-b8c5ccad8379" (UID: "17e34896-a3ff-4810-a21e-b8c5ccad8379"). InnerVolumeSpecName "kube-api-access-p9gd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.406018 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.412432 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.429582 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17e34896-a3ff-4810-a21e-b8c5ccad8379-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.429724 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9gd9\" (UniqueName: \"kubernetes.io/projected/17e34896-a3ff-4810-a21e-b8c5ccad8379-kube-api-access-p9gd9\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.614499 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb738854-dc39-46ff-b965-b69efbbec865" path="/var/lib/kubelet/pods/cb738854-dc39-46ff-b965-b69efbbec865/volumes" Oct 14 07:12:54 crc kubenswrapper[5018]: I1014 07:12:54.948911 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.074771 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-52nkw_17e34896-a3ff-4810-a21e-b8c5ccad8379/ovs-vswitchd/0.log" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.076328 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-52nkw" event={"ID":"17e34896-a3ff-4810-a21e-b8c5ccad8379","Type":"ContainerDied","Data":"2814725534bf0457611cad01eed2a88b8298cab45bb48662d892ff43122e2050"} Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.076387 5018 scope.go:117] "RemoveContainer" containerID="7d3952766ce431f04ff567c2e0ef933655ccba0395cfb7e85161c216e7d0e5c4" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.076649 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-52nkw" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.089858 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"50b722d0-93f3-41c5-a4e8-a314231f282b","Type":"ContainerDied","Data":"9317f296aa82a0881db11bdcabbe4cf06b4ce838519fad7b7e76b7ea26d99ee0"} Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.090012 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.113772 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.119404 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-52nkw"] Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.122034 5018 scope.go:117] "RemoveContainer" containerID="312a22f77e66c0d35fde7d32591f48b04cbf98597a52f6efbec5aec6ea0a933b" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.139341 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache\") pod \"50b722d0-93f3-41c5-a4e8-a314231f282b\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.139605 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"50b722d0-93f3-41c5-a4e8-a314231f282b\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.139943 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wn8r\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r\") pod \"50b722d0-93f3-41c5-a4e8-a314231f282b\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.140300 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") pod \"50b722d0-93f3-41c5-a4e8-a314231f282b\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.140566 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock\") pod \"50b722d0-93f3-41c5-a4e8-a314231f282b\" (UID: \"50b722d0-93f3-41c5-a4e8-a314231f282b\") " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.140991 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock" (OuterVolumeSpecName: "lock") pod "50b722d0-93f3-41c5-a4e8-a314231f282b" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.141255 5018 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-lock\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.141282 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache" (OuterVolumeSpecName: "cache") pod "50b722d0-93f3-41c5-a4e8-a314231f282b" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.144329 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r" (OuterVolumeSpecName: "kube-api-access-4wn8r") pod "50b722d0-93f3-41c5-a4e8-a314231f282b" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b"). InnerVolumeSpecName "kube-api-access-4wn8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.144505 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "swift") pod "50b722d0-93f3-41c5-a4e8-a314231f282b" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.146004 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "50b722d0-93f3-41c5-a4e8-a314231f282b" (UID: "50b722d0-93f3-41c5-a4e8-a314231f282b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.148799 5018 scope.go:117] "RemoveContainer" containerID="2c75064d547e5956672b603f9c1a06a4e5637ac16ca15b6e701e1c42346cfa93" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.243491 5018 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/50b722d0-93f3-41c5-a4e8-a314231f282b-cache\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.243561 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.243585 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wn8r\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-kube-api-access-4wn8r\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.243605 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/50b722d0-93f3-41c5-a4e8-a314231f282b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.243509 5018 scope.go:117] "RemoveContainer" containerID="f75caf8dbbd7c4743e32fa90f2af02023f5f583c5e27a58a4d2ae6216d5c09b9" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.263978 5018 scope.go:117] "RemoveContainer" containerID="7464ef303afd6563d9b11c50e58be39b3a6ce6d6bc0c340a249ad132b5dfdcba" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.271864 5018 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.290457 5018 scope.go:117] "RemoveContainer" containerID="969d5f8aadb42ee0593e5d86418c4fdb29ca61ce0f36d2a945eec3737e42b189" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.318156 5018 scope.go:117] "RemoveContainer" containerID="ec680e0a29290410e29c3d3731406e8216b29a7138c4bf28271899256ee732cc" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.344303 5018 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.349244 5018 scope.go:117] "RemoveContainer" containerID="599bb15e72a850b4e7c1ebc86b49f02ba6953b1cd2083f0f540b8f3b91ff08c0" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.374720 5018 scope.go:117] "RemoveContainer" containerID="963b3d72e03491fd9f8e960e77328e431e858c28e5183636d20edce5d6da90ec" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.399777 5018 scope.go:117] "RemoveContainer" containerID="1c34c44d438a3d6af5f9816098723851b45124c4a7a2d7e1641ec047b86764eb" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.417614 5018 scope.go:117] "RemoveContainer" containerID="b05e853fcda282db718ede400b9daeb449fd5e7cd0b75409d593e3b167779576" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.451992 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.459746 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.460268 5018 scope.go:117] "RemoveContainer" containerID="ebd9c6ab8467a65fedbce41b1ce78986ad397f0ca50597aaea953816e194054d" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.475905 5018 scope.go:117] "RemoveContainer" containerID="80c27ba55cfa5a682b1e2072de8e447382019d0b79b38cf3f4de1b47c35a429d" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.499195 5018 scope.go:117] "RemoveContainer" containerID="e58847c0dec122f46ec351ebf113ea26a71f6dd0002b59bf78203f30ea173dca" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.518309 5018 scope.go:117] "RemoveContainer" containerID="ee71c8af72a00afcf6ab9a1af606dec3a2657101b03f635f39d3c3a416113250" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.544430 5018 scope.go:117] "RemoveContainer" containerID="a9db9d72a3a3664bb31edfeb916708e025a2ea1bc87363e6693b7d4ae2caf69c" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.590546 5018 scope.go:117] "RemoveContainer" containerID="b83fc34a9d177c956e778ff75097967bb0fc448f9d4a508069b6032e5cd85729" Oct 14 07:12:55 crc kubenswrapper[5018]: I1014 07:12:55.610371 5018 scope.go:117] "RemoveContainer" containerID="422e0f3e0acbeb85eabbf4a53967dd8d157d1f9364ac749656bd718ee3f5b1e7" Oct 14 07:12:56 crc kubenswrapper[5018]: I1014 07:12:56.619677 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" path="/var/lib/kubelet/pods/17e34896-a3ff-4810-a21e-b8c5ccad8379/volumes" Oct 14 07:12:56 crc kubenswrapper[5018]: I1014 07:12:56.621368 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" path="/var/lib/kubelet/pods/50b722d0-93f3-41c5-a4e8-a314231f282b/volumes" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.728887 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hswfh"] Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729684 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="mysql-bootstrap" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729706 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="mysql-bootstrap" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729725 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="swift-recon-cron" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729739 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="swift-recon-cron" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729767 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729780 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729802 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729814 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729832 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" containerName="kube-state-metrics" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729845 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" containerName="kube-state-metrics" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729868 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729880 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729899 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="probe" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729910 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="probe" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729930 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" containerName="memcached" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729942 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" containerName="memcached" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729956 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729968 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-server" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.729986 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.729998 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730014 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ef95f6-1b6d-459a-b331-b114fb6205fb" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730026 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ef95f6-1b6d-459a-b331-b114fb6205fb" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730048 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730060 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730076 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="ovsdbserver-nb" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730087 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="ovsdbserver-nb" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730121 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730136 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730152 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="cinder-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730163 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="cinder-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730179 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="init" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730192 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="init" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730213 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730225 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730241 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730253 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730266 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730278 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730291 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="rsync" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730302 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="rsync" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730320 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730332 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730345 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730356 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730371 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730383 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730405 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0a409e-4a1f-471e-a5d5-b01213e5c0df" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730417 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0a409e-4a1f-471e-a5d5-b01213e5c0df" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730433 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730445 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730459 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730471 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730495 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-central-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730508 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-central-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730523 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="sg-core" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730535 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="sg-core" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730550 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730564 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730578 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730591 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730608 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f737b49-f2ba-4a7c-9b07-d934d5122363" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730648 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f737b49-f2ba-4a7c-9b07-d934d5122363" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730666 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730677 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730694 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730705 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730729 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730740 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-server" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730760 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730771 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730791 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730803 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-server" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730822 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730834 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730858 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d08361-7993-4757-afe3-1e9eeb5c1448" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730869 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d08361-7993-4757-afe3-1e9eeb5c1448" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730884 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730896 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730912 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730923 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730938 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" containerName="nova-cell1-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730950 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" containerName="nova-cell1-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.730970 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.730982 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731000 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerName="nova-scheduler-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731012 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerName="nova-scheduler-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731035 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-reaper" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731047 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-reaper" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731066 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731089 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731105 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731117 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731140 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server-init" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731152 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server-init" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731166 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731178 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731194 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-metadata" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731206 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-metadata" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731226 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731238 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731259 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-expirer" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731271 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-expirer" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731286 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731300 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731317 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731328 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731351 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2c0a409-0cc4-49a4-a1ad-3de144f90e49" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731363 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2c0a409-0cc4-49a4-a1ad-3de144f90e49" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731376 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731388 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-server" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731405 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="ovsdbserver-sb" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731417 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="ovsdbserver-sb" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731434 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731448 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731469 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731481 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731499 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731511 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731534 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="mysql-bootstrap" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731546 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="mysql-bootstrap" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731567 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerName="keystone-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731579 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerName="keystone-api" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731599 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731610 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731653 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731665 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731687 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="setup-container" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731699 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="setup-container" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731721 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a3ff62-8fcf-47aa-b080-b94042e137be" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731733 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a3ff62-8fcf-47aa-b080-b94042e137be" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731753 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731765 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731786 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-notification-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731798 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-notification-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731812 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731823 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731841 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731853 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731869 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" containerName="nova-cell0-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731882 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" containerName="nova-cell0-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731901 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731913 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731930 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="dnsmasq-dns" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731942 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="dnsmasq-dns" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731963 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="setup-container" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.731975 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="setup-container" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.731992 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732003 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.732016 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732027 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.732048 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6719ab90-8279-40bf-9544-b35118039b0e" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732059 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6719ab90-8279-40bf-9544-b35118039b0e" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: E1014 07:13:03.732078 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732090 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732318 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2c0a409-0cc4-49a4-a1ad-3de144f90e49" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732343 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-reaper" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732358 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732374 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732393 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="467829ec-f0b0-490c-80fb-25dc08df07b4" containerName="nova-scheduler-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732408 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732428 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca7068c-e306-467d-972d-6ec8439d2b85" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732440 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e9ef0b-92f8-43fc-8b2f-e67ab1eabd4e" containerName="memcached" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732453 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="swift-recon-cron" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732476 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="ovsdbserver-nb" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732497 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="ovsdbserver-sb" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732515 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732530 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732543 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732565 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a49d558a-c4f6-49d3-8210-c38799ab89c4" containerName="barbican-keystone-listener-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732586 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732604 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2372cde-de88-46c2-abc8-0bd1a9e7b961" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732649 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732671 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="63753948-7b37-4eeb-a378-2947a6caa55c" containerName="nova-cell1-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732690 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3673da9-258f-4052-916a-90eb07abe4e6" containerName="cinder-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732702 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732723 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732737 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732762 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-expirer" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732782 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="ovn-northd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732796 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc86ed1d-ae09-4120-8e15-2a960334a853" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732815 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c736e3c0-1c92-463c-aa8e-104d6bb2c6c0" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732830 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0a409e-4a1f-471e-a5d5-b01213e5c0df" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732849 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b076446-7046-4b7a-b315-f8a560d5604b" containerName="nova-cell0-conductor-conductor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732867 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa6d249-3441-4d16-9a74-a5232861aa21" containerName="galera" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732896 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d08361-7993-4757-afe3-1e9eeb5c1448" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732913 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732929 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="248e38fb-736b-4333-a184-5a6fdd04daeb" containerName="nova-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732945 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732962 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dae2a92-e082-4271-b517-2f9219a07f2b" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732979 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cced7bf3-2f55-4c00-87f0-b7e17c53ee0e" containerName="proxy-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.732994 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-central-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733014 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ddcde8d-45fd-419b-8751-18db89bf729d" containerName="rabbitmq" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733029 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab012259-4546-400d-9e34-217dea0ee019" containerName="placement-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733046 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="53392f65-7a1b-49fc-96a4-341f237e9288" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733059 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ef95f6-1b6d-459a-b331-b114fb6205fb" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733071 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ebfec5-253f-4fd8-88f5-52f38b7f2e2a" containerName="barbican-worker-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733088 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="390ed4c4-a917-47a0-83e5-5c05fbaf2005" containerName="keystone-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733108 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b77f89-1956-43e0-b6df-dd4d390e4cef" containerName="barbican-api-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733135 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovsdb-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733152 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc288e96-07a9-43b9-b153-a9246dd704eb" containerName="dnsmasq-dns" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733166 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e34896-a3ff-4810-a21e-b8c5ccad8379" containerName="ovs-vswitchd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733181 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733195 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-replicator" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733209 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a69b66-5d0d-410b-9d2e-6179c8079077" containerName="nova-metadata-metadata" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733222 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1751d1e2-6a95-4a5f-b588-a58bffb94896" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733242 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-log" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733255 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="sg-core" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733271 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="probe" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733285 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a3ff62-8fcf-47aa-b080-b94042e137be" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733299 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9387da-03d5-46e4-b442-aecb5fd1e766" containerName="ceilometer-notification-agent" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733313 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6719ab90-8279-40bf-9544-b35118039b0e" containerName="mariadb-account-delete" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733332 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733345 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="rsync" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733357 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="52261d3b-8ca7-4632-9f05-a8e5b656f397" containerName="glance-httpd" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733374 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd8d742-9217-4dd6-a035-49fab332a574" containerName="ovn-controller" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733385 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f737b49-f2ba-4a7c-9b07-d934d5122363" containerName="openstack-network-exporter" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733404 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="account-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733425 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733438 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733449 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aa3b2a-b639-4e29-9593-0da33b053fbd" containerName="neutron-api" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733469 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb738854-dc39-46ff-b965-b69efbbec865" containerName="cinder-scheduler" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733487 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="container-server" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733500 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-auditor" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733513 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b722d0-93f3-41c5-a4e8-a314231f282b" containerName="object-updater" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.733534 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="66a64e8d-ccfb-4f2c-8ee8-b509d7ec4bbe" containerName="kube-state-metrics" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.735897 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.747989 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hswfh"] Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.783504 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wv5\" (UniqueName: \"kubernetes.io/projected/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-kube-api-access-d5wv5\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.783563 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-catalog-content\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.783638 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-utilities\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.884992 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wv5\" (UniqueName: \"kubernetes.io/projected/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-kube-api-access-d5wv5\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.885051 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-catalog-content\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.885106 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-utilities\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.885740 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-utilities\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.886432 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-catalog-content\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:03 crc kubenswrapper[5018]: I1014 07:13:03.908234 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wv5\" (UniqueName: \"kubernetes.io/projected/d5c73ed3-ccf0-48fb-9edf-93748b24edf7-kube-api-access-d5wv5\") pod \"redhat-operators-hswfh\" (UID: \"d5c73ed3-ccf0-48fb-9edf-93748b24edf7\") " pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:04 crc kubenswrapper[5018]: I1014 07:13:04.068212 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:04 crc kubenswrapper[5018]: I1014 07:13:04.515718 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hswfh"] Oct 14 07:13:05 crc kubenswrapper[5018]: I1014 07:13:05.213427 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5c73ed3-ccf0-48fb-9edf-93748b24edf7" containerID="c705529d57c035a74bd76a5502b78f9ada488384c5fb000165e7aa50b26db002" exitCode=0 Oct 14 07:13:05 crc kubenswrapper[5018]: I1014 07:13:05.213533 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hswfh" event={"ID":"d5c73ed3-ccf0-48fb-9edf-93748b24edf7","Type":"ContainerDied","Data":"c705529d57c035a74bd76a5502b78f9ada488384c5fb000165e7aa50b26db002"} Oct 14 07:13:05 crc kubenswrapper[5018]: I1014 07:13:05.214058 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hswfh" event={"ID":"d5c73ed3-ccf0-48fb-9edf-93748b24edf7","Type":"ContainerStarted","Data":"547907ecd2df516c3d314bc17f1c32df015d72a1efe75f052b65881d3ad2c179"} Oct 14 07:13:05 crc kubenswrapper[5018]: I1014 07:13:05.217295 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:13:12 crc kubenswrapper[5018]: I1014 07:13:12.273871 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hswfh" event={"ID":"d5c73ed3-ccf0-48fb-9edf-93748b24edf7","Type":"ContainerStarted","Data":"2680aff1ce0cd25218c77bee3bdbede2b81b83fa2fe81dda7dbe6ac544e9c829"} Oct 14 07:13:13 crc kubenswrapper[5018]: I1014 07:13:13.293149 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5c73ed3-ccf0-48fb-9edf-93748b24edf7" containerID="2680aff1ce0cd25218c77bee3bdbede2b81b83fa2fe81dda7dbe6ac544e9c829" exitCode=0 Oct 14 07:13:13 crc kubenswrapper[5018]: I1014 07:13:13.293273 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hswfh" event={"ID":"d5c73ed3-ccf0-48fb-9edf-93748b24edf7","Type":"ContainerDied","Data":"2680aff1ce0cd25218c77bee3bdbede2b81b83fa2fe81dda7dbe6ac544e9c829"} Oct 14 07:13:14 crc kubenswrapper[5018]: I1014 07:13:14.306688 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hswfh" event={"ID":"d5c73ed3-ccf0-48fb-9edf-93748b24edf7","Type":"ContainerStarted","Data":"661e28b3b421f3b93b6e2d05f88aa2d0568eefeca25b1d2bfb170adce4f21046"} Oct 14 07:13:14 crc kubenswrapper[5018]: I1014 07:13:14.334451 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hswfh" podStartSLOduration=2.555338056 podStartE2EDuration="11.334424621s" podCreationTimestamp="2025-10-14 07:13:03 +0000 UTC" firstStartedPulling="2025-10-14 07:13:05.216666868 +0000 UTC m=+1401.800713535" lastFinishedPulling="2025-10-14 07:13:13.995753443 +0000 UTC m=+1410.579800100" observedRunningTime="2025-10-14 07:13:14.333071473 +0000 UTC m=+1410.917118130" watchObservedRunningTime="2025-10-14 07:13:14.334424621 +0000 UTC m=+1410.918471288" Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.069247 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.071383 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.142595 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.490530 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hswfh" Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.598606 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hswfh"] Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.654460 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 07:13:24 crc kubenswrapper[5018]: I1014 07:13:24.655208 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-db6sr" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="registry-server" containerID="cri-o://d988f06ec7750d8716be5da2d3fd09f55f000182661cb82375fd7e5084a03133" gracePeriod=2 Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.421463 5018 generic.go:334] "Generic (PLEG): container finished" podID="be2fff80-efbf-4372-a47c-259e5a502e33" containerID="d988f06ec7750d8716be5da2d3fd09f55f000182661cb82375fd7e5084a03133" exitCode=0 Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.421559 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerDied","Data":"d988f06ec7750d8716be5da2d3fd09f55f000182661cb82375fd7e5084a03133"} Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.779596 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.931896 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content\") pod \"be2fff80-efbf-4372-a47c-259e5a502e33\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.932012 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b965f\" (UniqueName: \"kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f\") pod \"be2fff80-efbf-4372-a47c-259e5a502e33\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.932036 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities\") pod \"be2fff80-efbf-4372-a47c-259e5a502e33\" (UID: \"be2fff80-efbf-4372-a47c-259e5a502e33\") " Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.932750 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities" (OuterVolumeSpecName: "utilities") pod "be2fff80-efbf-4372-a47c-259e5a502e33" (UID: "be2fff80-efbf-4372-a47c-259e5a502e33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:13:25 crc kubenswrapper[5018]: I1014 07:13:25.936863 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f" (OuterVolumeSpecName: "kube-api-access-b965f") pod "be2fff80-efbf-4372-a47c-259e5a502e33" (UID: "be2fff80-efbf-4372-a47c-259e5a502e33"). InnerVolumeSpecName "kube-api-access-b965f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.006830 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be2fff80-efbf-4372-a47c-259e5a502e33" (UID: "be2fff80-efbf-4372-a47c-259e5a502e33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.033383 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.033416 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b965f\" (UniqueName: \"kubernetes.io/projected/be2fff80-efbf-4372-a47c-259e5a502e33-kube-api-access-b965f\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.033432 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fff80-efbf-4372-a47c-259e5a502e33-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.432658 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db6sr" event={"ID":"be2fff80-efbf-4372-a47c-259e5a502e33","Type":"ContainerDied","Data":"26516a7ac17da83a531e73975130dfd12c65a8ee66b7c123de35be8e401b7e30"} Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.432735 5018 scope.go:117] "RemoveContainer" containerID="d988f06ec7750d8716be5da2d3fd09f55f000182661cb82375fd7e5084a03133" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.432685 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db6sr" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.457605 5018 scope.go:117] "RemoveContainer" containerID="19af5e219d45cf00b4a302a66765b247cd210f21f36e2a2cef71c054cbfb9252" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.470071 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.474867 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-db6sr"] Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.487112 5018 scope.go:117] "RemoveContainer" containerID="8ac45eb863dcb797a277ad4ce6b7debb115c1b943fc32c4c9451123e5824e7b9" Oct 14 07:13:26 crc kubenswrapper[5018]: I1014 07:13:26.614796 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" path="/var/lib/kubelet/pods/be2fff80-efbf-4372-a47c-259e5a502e33/volumes" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.583762 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:13:53 crc kubenswrapper[5018]: E1014 07:13:53.584939 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="registry-server" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.584961 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="registry-server" Oct 14 07:13:53 crc kubenswrapper[5018]: E1014 07:13:53.584995 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="extract-utilities" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.585009 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="extract-utilities" Oct 14 07:13:53 crc kubenswrapper[5018]: E1014 07:13:53.585058 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="extract-content" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.585069 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="extract-content" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.585307 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="be2fff80-efbf-4372-a47c-259e5a502e33" containerName="registry-server" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.586898 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.592401 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nttck\" (UniqueName: \"kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.592546 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.592741 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.609435 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.693611 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.694530 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nttck\" (UniqueName: \"kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.694602 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.695018 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.695289 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.718823 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nttck\" (UniqueName: \"kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck\") pod \"certified-operators-fhp6j\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:53 crc kubenswrapper[5018]: I1014 07:13:53.915283 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:13:54 crc kubenswrapper[5018]: I1014 07:13:54.422847 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:13:54 crc kubenswrapper[5018]: W1014 07:13:54.431076 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod731c4a2d_c394_409f_b6b8_79803cb72c3b.slice/crio-2d4da55c34affde07d53734c139ac08b986bd855286af567e82405b688e81715 WatchSource:0}: Error finding container 2d4da55c34affde07d53734c139ac08b986bd855286af567e82405b688e81715: Status 404 returned error can't find the container with id 2d4da55c34affde07d53734c139ac08b986bd855286af567e82405b688e81715 Oct 14 07:13:54 crc kubenswrapper[5018]: I1014 07:13:54.754674 5018 generic.go:334] "Generic (PLEG): container finished" podID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerID="52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca" exitCode=0 Oct 14 07:13:54 crc kubenswrapper[5018]: I1014 07:13:54.755048 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerDied","Data":"52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca"} Oct 14 07:13:54 crc kubenswrapper[5018]: I1014 07:13:54.756608 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerStarted","Data":"2d4da55c34affde07d53734c139ac08b986bd855286af567e82405b688e81715"} Oct 14 07:13:55 crc kubenswrapper[5018]: I1014 07:13:55.771217 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerStarted","Data":"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77"} Oct 14 07:13:56 crc kubenswrapper[5018]: I1014 07:13:56.784489 5018 generic.go:334] "Generic (PLEG): container finished" podID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerID="f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77" exitCode=0 Oct 14 07:13:56 crc kubenswrapper[5018]: I1014 07:13:56.784599 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerDied","Data":"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77"} Oct 14 07:13:56 crc kubenswrapper[5018]: I1014 07:13:56.784907 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerStarted","Data":"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13"} Oct 14 07:13:56 crc kubenswrapper[5018]: I1014 07:13:56.815856 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fhp6j" podStartSLOduration=2.32027239 podStartE2EDuration="3.815830272s" podCreationTimestamp="2025-10-14 07:13:53 +0000 UTC" firstStartedPulling="2025-10-14 07:13:54.758863164 +0000 UTC m=+1451.342909831" lastFinishedPulling="2025-10-14 07:13:56.254421056 +0000 UTC m=+1452.838467713" observedRunningTime="2025-10-14 07:13:56.809659785 +0000 UTC m=+1453.393706472" watchObservedRunningTime="2025-10-14 07:13:56.815830272 +0000 UTC m=+1453.399876939" Oct 14 07:14:01 crc kubenswrapper[5018]: I1014 07:14:01.918376 5018 scope.go:117] "RemoveContainer" containerID="1889b1037262e6670ececf5f3cc0fad455f0f676c091666af0087aa6c4a7318a" Oct 14 07:14:01 crc kubenswrapper[5018]: I1014 07:14:01.962865 5018 scope.go:117] "RemoveContainer" containerID="0090c0a99892af1c98856d4013a796c1ad868a038ad76ec6b002625b0cb74e54" Oct 14 07:14:02 crc kubenswrapper[5018]: I1014 07:14:02.007368 5018 scope.go:117] "RemoveContainer" containerID="72aaf2f3ea790510bd282622e3cc4b3f450e19ea9c51089f9720e414ce4c5e79" Oct 14 07:14:02 crc kubenswrapper[5018]: I1014 07:14:02.036740 5018 scope.go:117] "RemoveContainer" containerID="295bcb9c9f247ccad615ca5131bff760f8822e3791874a3fa3874a7d346939e7" Oct 14 07:14:02 crc kubenswrapper[5018]: I1014 07:14:02.070750 5018 scope.go:117] "RemoveContainer" containerID="f106d8478ac0f489c8ee548d9ecffc0f6885c004e9abab5c884eddda23eedfea" Oct 14 07:14:02 crc kubenswrapper[5018]: I1014 07:14:02.101573 5018 scope.go:117] "RemoveContainer" containerID="2857b484232f0cbf1617641f754477a7d6b6fa7212de8ab2bc4d14513764179d" Oct 14 07:14:02 crc kubenswrapper[5018]: I1014 07:14:02.130307 5018 scope.go:117] "RemoveContainer" containerID="553347b1b80d8f4b315497ba5161f42a2b187bace35cb9dba665dc45ea96194d" Oct 14 07:14:03 crc kubenswrapper[5018]: I1014 07:14:03.915902 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:03 crc kubenswrapper[5018]: I1014 07:14:03.917251 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:03 crc kubenswrapper[5018]: I1014 07:14:03.977287 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.245589 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.249022 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.256239 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.365321 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.365552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hwg2\" (UniqueName: \"kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.365663 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.466824 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hwg2\" (UniqueName: \"kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.466907 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.467052 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.467594 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.467798 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.500323 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hwg2\" (UniqueName: \"kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2\") pod \"community-operators-v5rdg\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.581308 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:04 crc kubenswrapper[5018]: I1014 07:14:04.963901 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:05 crc kubenswrapper[5018]: I1014 07:14:05.064497 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:05 crc kubenswrapper[5018]: I1014 07:14:05.907723 5018 generic.go:334] "Generic (PLEG): container finished" podID="b5fd98c7-a558-456d-989c-2113545a9023" containerID="cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1" exitCode=0 Oct 14 07:14:05 crc kubenswrapper[5018]: I1014 07:14:05.907820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerDied","Data":"cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1"} Oct 14 07:14:05 crc kubenswrapper[5018]: I1014 07:14:05.908232 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerStarted","Data":"c6b7f8d037ea807225339fe15e73b98b20639e4d21345845c3608a2741bd671b"} Oct 14 07:14:07 crc kubenswrapper[5018]: I1014 07:14:07.223658 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:14:07 crc kubenswrapper[5018]: I1014 07:14:07.935135 5018 generic.go:334] "Generic (PLEG): container finished" podID="b5fd98c7-a558-456d-989c-2113545a9023" containerID="4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3" exitCode=0 Oct 14 07:14:07 crc kubenswrapper[5018]: I1014 07:14:07.935297 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerDied","Data":"4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3"} Oct 14 07:14:07 crc kubenswrapper[5018]: I1014 07:14:07.935750 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fhp6j" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="registry-server" containerID="cri-o://f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13" gracePeriod=2 Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.427975 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.532525 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content\") pod \"731c4a2d-c394-409f-b6b8-79803cb72c3b\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.532859 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities\") pod \"731c4a2d-c394-409f-b6b8-79803cb72c3b\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.533010 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nttck\" (UniqueName: \"kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck\") pod \"731c4a2d-c394-409f-b6b8-79803cb72c3b\" (UID: \"731c4a2d-c394-409f-b6b8-79803cb72c3b\") " Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.533495 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities" (OuterVolumeSpecName: "utilities") pod "731c4a2d-c394-409f-b6b8-79803cb72c3b" (UID: "731c4a2d-c394-409f-b6b8-79803cb72c3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.544340 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck" (OuterVolumeSpecName: "kube-api-access-nttck") pod "731c4a2d-c394-409f-b6b8-79803cb72c3b" (UID: "731c4a2d-c394-409f-b6b8-79803cb72c3b"). InnerVolumeSpecName "kube-api-access-nttck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.597692 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "731c4a2d-c394-409f-b6b8-79803cb72c3b" (UID: "731c4a2d-c394-409f-b6b8-79803cb72c3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.634672 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.634702 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nttck\" (UniqueName: \"kubernetes.io/projected/731c4a2d-c394-409f-b6b8-79803cb72c3b-kube-api-access-nttck\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.634713 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731c4a2d-c394-409f-b6b8-79803cb72c3b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.945368 5018 generic.go:334] "Generic (PLEG): container finished" podID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerID="f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13" exitCode=0 Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.945443 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fhp6j" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.945452 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerDied","Data":"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13"} Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.945483 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fhp6j" event={"ID":"731c4a2d-c394-409f-b6b8-79803cb72c3b","Type":"ContainerDied","Data":"2d4da55c34affde07d53734c139ac08b986bd855286af567e82405b688e81715"} Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.945502 5018 scope.go:117] "RemoveContainer" containerID="f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.947735 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerStarted","Data":"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3"} Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.968151 5018 scope.go:117] "RemoveContainer" containerID="f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77" Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.968643 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.974966 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fhp6j"] Oct 14 07:14:08 crc kubenswrapper[5018]: I1014 07:14:08.993017 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v5rdg" podStartSLOduration=2.479143022 podStartE2EDuration="4.992997446s" podCreationTimestamp="2025-10-14 07:14:04 +0000 UTC" firstStartedPulling="2025-10-14 07:14:05.910858828 +0000 UTC m=+1462.494905455" lastFinishedPulling="2025-10-14 07:14:08.424713242 +0000 UTC m=+1465.008759879" observedRunningTime="2025-10-14 07:14:08.985708548 +0000 UTC m=+1465.569755195" watchObservedRunningTime="2025-10-14 07:14:08.992997446 +0000 UTC m=+1465.577044083" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.001075 5018 scope.go:117] "RemoveContainer" containerID="52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.028528 5018 scope.go:117] "RemoveContainer" containerID="f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13" Oct 14 07:14:09 crc kubenswrapper[5018]: E1014 07:14:09.029089 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13\": container with ID starting with f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13 not found: ID does not exist" containerID="f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.029124 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13"} err="failed to get container status \"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13\": rpc error: code = NotFound desc = could not find container \"f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13\": container with ID starting with f51b4c76390bbdd765ba0b1461997a27580fedd73502fd8e01b7007c2da1ad13 not found: ID does not exist" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.029148 5018 scope.go:117] "RemoveContainer" containerID="f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77" Oct 14 07:14:09 crc kubenswrapper[5018]: E1014 07:14:09.029432 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77\": container with ID starting with f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77 not found: ID does not exist" containerID="f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.029502 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77"} err="failed to get container status \"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77\": rpc error: code = NotFound desc = could not find container \"f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77\": container with ID starting with f4f9c397e1640d62e16b30afc4c7e452ec4b905589ed3f69e2a8607f91356e77 not found: ID does not exist" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.029537 5018 scope.go:117] "RemoveContainer" containerID="52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca" Oct 14 07:14:09 crc kubenswrapper[5018]: E1014 07:14:09.030025 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca\": container with ID starting with 52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca not found: ID does not exist" containerID="52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca" Oct 14 07:14:09 crc kubenswrapper[5018]: I1014 07:14:09.030069 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca"} err="failed to get container status \"52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca\": rpc error: code = NotFound desc = could not find container \"52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca\": container with ID starting with 52cb8edf42c618245f9fe869796813db83749737ba5ca4af088ea05f3a1cf2ca not found: ID does not exist" Oct 14 07:14:10 crc kubenswrapper[5018]: I1014 07:14:10.614881 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" path="/var/lib/kubelet/pods/731c4a2d-c394-409f-b6b8-79803cb72c3b/volumes" Oct 14 07:14:14 crc kubenswrapper[5018]: I1014 07:14:14.581503 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:14 crc kubenswrapper[5018]: I1014 07:14:14.582946 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:14 crc kubenswrapper[5018]: I1014 07:14:14.655018 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:15 crc kubenswrapper[5018]: I1014 07:14:15.127305 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:15 crc kubenswrapper[5018]: I1014 07:14:15.197674 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.025410 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v5rdg" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="registry-server" containerID="cri-o://3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3" gracePeriod=2 Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.482164 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.569743 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hwg2\" (UniqueName: \"kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2\") pod \"b5fd98c7-a558-456d-989c-2113545a9023\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.569795 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content\") pod \"b5fd98c7-a558-456d-989c-2113545a9023\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.569880 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities\") pod \"b5fd98c7-a558-456d-989c-2113545a9023\" (UID: \"b5fd98c7-a558-456d-989c-2113545a9023\") " Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.570787 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities" (OuterVolumeSpecName: "utilities") pod "b5fd98c7-a558-456d-989c-2113545a9023" (UID: "b5fd98c7-a558-456d-989c-2113545a9023"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.578626 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2" (OuterVolumeSpecName: "kube-api-access-4hwg2") pod "b5fd98c7-a558-456d-989c-2113545a9023" (UID: "b5fd98c7-a558-456d-989c-2113545a9023"). InnerVolumeSpecName "kube-api-access-4hwg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.638957 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5fd98c7-a558-456d-989c-2113545a9023" (UID: "b5fd98c7-a558-456d-989c-2113545a9023"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.672079 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hwg2\" (UniqueName: \"kubernetes.io/projected/b5fd98c7-a558-456d-989c-2113545a9023-kube-api-access-4hwg2\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.672117 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:17 crc kubenswrapper[5018]: I1014 07:14:17.672126 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fd98c7-a558-456d-989c-2113545a9023-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.039367 5018 generic.go:334] "Generic (PLEG): container finished" podID="b5fd98c7-a558-456d-989c-2113545a9023" containerID="3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3" exitCode=0 Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.039432 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerDied","Data":"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3"} Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.039473 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v5rdg" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.039499 5018 scope.go:117] "RemoveContainer" containerID="3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.039482 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v5rdg" event={"ID":"b5fd98c7-a558-456d-989c-2113545a9023","Type":"ContainerDied","Data":"c6b7f8d037ea807225339fe15e73b98b20639e4d21345845c3608a2741bd671b"} Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.078248 5018 scope.go:117] "RemoveContainer" containerID="4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.082262 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.093337 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v5rdg"] Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.118401 5018 scope.go:117] "RemoveContainer" containerID="cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.139145 5018 scope.go:117] "RemoveContainer" containerID="3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3" Oct 14 07:14:18 crc kubenswrapper[5018]: E1014 07:14:18.139691 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3\": container with ID starting with 3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3 not found: ID does not exist" containerID="3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.139749 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3"} err="failed to get container status \"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3\": rpc error: code = NotFound desc = could not find container \"3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3\": container with ID starting with 3fa7e24a5d8250f9e73b5eaf617afeb35575b16ef14a6efbc6d7349e6f8926a3 not found: ID does not exist" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.139779 5018 scope.go:117] "RemoveContainer" containerID="4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3" Oct 14 07:14:18 crc kubenswrapper[5018]: E1014 07:14:18.140116 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3\": container with ID starting with 4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3 not found: ID does not exist" containerID="4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.140219 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3"} err="failed to get container status \"4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3\": rpc error: code = NotFound desc = could not find container \"4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3\": container with ID starting with 4dd74321ef8df727d8b4c5244cda5ad3773ad086c76e78c4977b10bb375578e3 not found: ID does not exist" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.140314 5018 scope.go:117] "RemoveContainer" containerID="cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1" Oct 14 07:14:18 crc kubenswrapper[5018]: E1014 07:14:18.140600 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1\": container with ID starting with cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1 not found: ID does not exist" containerID="cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.140687 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1"} err="failed to get container status \"cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1\": rpc error: code = NotFound desc = could not find container \"cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1\": container with ID starting with cf129b74bdf3a1a15a3a2bb43ea21d2ea3cb59c2a944eb32645678fbd859d0f1 not found: ID does not exist" Oct 14 07:14:18 crc kubenswrapper[5018]: I1014 07:14:18.614695 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fd98c7-a558-456d-989c-2113545a9023" path="/var/lib/kubelet/pods/b5fd98c7-a558-456d-989c-2113545a9023/volumes" Oct 14 07:14:32 crc kubenswrapper[5018]: I1014 07:14:32.463786 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:14:32 crc kubenswrapper[5018]: I1014 07:14:32.464540 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.155972 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn"] Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158015 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="extract-utilities" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.158124 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="extract-utilities" Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158217 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="extract-content" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.158330 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="extract-content" Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158428 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="extract-content" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.158507 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="extract-content" Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158597 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="extract-utilities" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.158703 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="extract-utilities" Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158792 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.158879 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: E1014 07:15:00.158997 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.159078 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.159338 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="731c4a2d-c394-409f-b6b8-79803cb72c3b" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.159436 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fd98c7-a558-456d-989c-2113545a9023" containerName="registry-server" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.160278 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.162318 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.162519 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.166109 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn"] Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.244228 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.244313 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.244439 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwwj\" (UniqueName: \"kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.346247 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwwj\" (UniqueName: \"kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.346344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.346493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.348503 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.357044 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.369684 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwwj\" (UniqueName: \"kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj\") pod \"collect-profiles-29340435-s8jtn\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.489502 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:00 crc kubenswrapper[5018]: I1014 07:15:00.697993 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn"] Oct 14 07:15:01 crc kubenswrapper[5018]: I1014 07:15:01.474345 5018 generic.go:334] "Generic (PLEG): container finished" podID="e7954f62-3572-426b-b442-bccaf1ed9287" containerID="f26836cf5d46f92d66a1bc414257d46795721b947cf5dbd9e888016bf3d1f3c0" exitCode=0 Oct 14 07:15:01 crc kubenswrapper[5018]: I1014 07:15:01.474464 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" event={"ID":"e7954f62-3572-426b-b442-bccaf1ed9287","Type":"ContainerDied","Data":"f26836cf5d46f92d66a1bc414257d46795721b947cf5dbd9e888016bf3d1f3c0"} Oct 14 07:15:01 crc kubenswrapper[5018]: I1014 07:15:01.474769 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" event={"ID":"e7954f62-3572-426b-b442-bccaf1ed9287","Type":"ContainerStarted","Data":"ed1743a49af38a33c6f20e2a8fddb1245e624578b6ec584be3d01382b6fa052c"} Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.327251 5018 scope.go:117] "RemoveContainer" containerID="9051ceab7493cbb60404bd9f62f8db4f133e15225ee6817015967faf21cef4f7" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.437796 5018 scope.go:117] "RemoveContainer" containerID="14bd600b1cdf54ac9e825fe0357c58d8425a49107b37b3119e48f3bee9fb6c6a" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.463711 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.463796 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.526585 5018 scope.go:117] "RemoveContainer" containerID="89ca651abd56061b3d4f4730fa92495f37b9b76c796d8b6fdf496276d6c696c2" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.546107 5018 scope.go:117] "RemoveContainer" containerID="a69669374cda679bf77ce540f42dd59f2d251b3623db67fb8c294ce9d28fd5b8" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.605826 5018 scope.go:117] "RemoveContainer" containerID="b310644847b9e57cc0725838b5a8468110e9a5e7031ec226076e90392203de20" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.634349 5018 scope.go:117] "RemoveContainer" containerID="75818e8b9032462b8d1fa44368e8a77b40f4d9c29e54cb0b2af0d3c3b89438b7" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.669773 5018 scope.go:117] "RemoveContainer" containerID="855cef651e93fc582bf4dfce02cadbb1a3d709865a3fde078f280d5e771dc195" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.687313 5018 scope.go:117] "RemoveContainer" containerID="1a0f3efdf0f6304ed75d79e70ae3f3de5dbdbc727d01370fd2c7e00dba266640" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.718669 5018 scope.go:117] "RemoveContainer" containerID="9e8038bd5aa84b2d30b0819224bdfdd25f526085072462f8308583a788feaba3" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.729239 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.746694 5018 scope.go:117] "RemoveContainer" containerID="1be68f6f0cba15c3bc66c45d56c706fc1f7a4a7231a6283e04c05aaac90e6232" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.769745 5018 scope.go:117] "RemoveContainer" containerID="c60820610e6717de5af409b5508e7a1c494ce74fdb1fcce6f6fbbec19e7ed673" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.776643 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume\") pod \"e7954f62-3572-426b-b442-bccaf1ed9287\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.776805 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume\") pod \"e7954f62-3572-426b-b442-bccaf1ed9287\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.776880 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrwwj\" (UniqueName: \"kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj\") pod \"e7954f62-3572-426b-b442-bccaf1ed9287\" (UID: \"e7954f62-3572-426b-b442-bccaf1ed9287\") " Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.777563 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7954f62-3572-426b-b442-bccaf1ed9287" (UID: "e7954f62-3572-426b-b442-bccaf1ed9287"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.782764 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7954f62-3572-426b-b442-bccaf1ed9287" (UID: "e7954f62-3572-426b-b442-bccaf1ed9287"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.784857 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj" (OuterVolumeSpecName: "kube-api-access-hrwwj") pod "e7954f62-3572-426b-b442-bccaf1ed9287" (UID: "e7954f62-3572-426b-b442-bccaf1ed9287"). InnerVolumeSpecName "kube-api-access-hrwwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.801905 5018 scope.go:117] "RemoveContainer" containerID="eb63bc0308d56dec44abe934bec14d642df71515d6ac71e9bf8281b994bb49a4" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.828826 5018 scope.go:117] "RemoveContainer" containerID="a76b17433cbbe67cc02ee352389554f84d8ac0be588c5e00e58f7115928d5574" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.853973 5018 scope.go:117] "RemoveContainer" containerID="9d51216aaee8d1ff3d268268f6c176157e2126998ecf32f4f13de3c851aac61f" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.878565 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7954f62-3572-426b-b442-bccaf1ed9287-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.878594 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrwwj\" (UniqueName: \"kubernetes.io/projected/e7954f62-3572-426b-b442-bccaf1ed9287-kube-api-access-hrwwj\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:02 crc kubenswrapper[5018]: I1014 07:15:02.878605 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7954f62-3572-426b-b442-bccaf1ed9287-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:03 crc kubenswrapper[5018]: I1014 07:15:03.491742 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" event={"ID":"e7954f62-3572-426b-b442-bccaf1ed9287","Type":"ContainerDied","Data":"ed1743a49af38a33c6f20e2a8fddb1245e624578b6ec584be3d01382b6fa052c"} Oct 14 07:15:03 crc kubenswrapper[5018]: I1014 07:15:03.491779 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed1743a49af38a33c6f20e2a8fddb1245e624578b6ec584be3d01382b6fa052c" Oct 14 07:15:03 crc kubenswrapper[5018]: I1014 07:15:03.491807 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.463659 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.464323 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.464382 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.465293 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.465387 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" gracePeriod=600 Oct 14 07:15:32 crc kubenswrapper[5018]: E1014 07:15:32.594597 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.841280 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" exitCode=0 Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.841350 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb"} Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.841416 5018 scope.go:117] "RemoveContainer" containerID="613260043e4d0e2bc9178e8980e7e18f8853cd6ae68881c011659d6fa5e9258d" Oct 14 07:15:32 crc kubenswrapper[5018]: I1014 07:15:32.842102 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:15:32 crc kubenswrapper[5018]: E1014 07:15:32.842573 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:15:46 crc kubenswrapper[5018]: I1014 07:15:46.605404 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:15:46 crc kubenswrapper[5018]: E1014 07:15:46.606586 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:16:01 crc kubenswrapper[5018]: I1014 07:16:01.605526 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:16:01 crc kubenswrapper[5018]: E1014 07:16:01.606299 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.047468 5018 scope.go:117] "RemoveContainer" containerID="9903204a108ca41d3c2402336f8b8ea88b9a08b94770a7155c078e2492dacc43" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.074024 5018 scope.go:117] "RemoveContainer" containerID="5f58d093c2e1b4df7035a9b6e39bb45dfdc1616b93fda503854192c6a3ac25c4" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.127991 5018 scope.go:117] "RemoveContainer" containerID="19c6afa7b1ea3758086293b2bae47d5f7aba7eaf8fffdb5f13115ab7520de515" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.185562 5018 scope.go:117] "RemoveContainer" containerID="eafbe1d749e70aac5a62d1ff29e85ed71cd319d9a0ecfadcdd4aec4654df7467" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.208721 5018 scope.go:117] "RemoveContainer" containerID="44ea1af93d522ee762df6bcf89d1c6d4e233fb449c6c50287d54ddf3ebb98aac" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.229754 5018 scope.go:117] "RemoveContainer" containerID="91711faedba7fc88d95405468244ef504b36f759b6a39014eb7510244825fb42" Oct 14 07:16:03 crc kubenswrapper[5018]: I1014 07:16:03.253083 5018 scope.go:117] "RemoveContainer" containerID="357f6156d3b1a125fd854af5001fabf852e7fd7d6d4c24cd609e0733b9bb922b" Oct 14 07:16:16 crc kubenswrapper[5018]: I1014 07:16:16.605532 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:16:16 crc kubenswrapper[5018]: E1014 07:16:16.607059 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:16:27 crc kubenswrapper[5018]: I1014 07:16:27.606808 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:16:27 crc kubenswrapper[5018]: E1014 07:16:27.607814 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.575922 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:31 crc kubenswrapper[5018]: E1014 07:16:31.576457 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7954f62-3572-426b-b442-bccaf1ed9287" containerName="collect-profiles" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.576469 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7954f62-3572-426b-b442-bccaf1ed9287" containerName="collect-profiles" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.576677 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7954f62-3572-426b-b442-bccaf1ed9287" containerName="collect-profiles" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.577570 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.596382 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.740788 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgxk9\" (UniqueName: \"kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.740871 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.741018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.842270 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.842351 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgxk9\" (UniqueName: \"kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.842394 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.842901 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.842932 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.867814 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgxk9\" (UniqueName: \"kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9\") pod \"redhat-marketplace-86hmd\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:31 crc kubenswrapper[5018]: I1014 07:16:31.941466 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:32 crc kubenswrapper[5018]: I1014 07:16:32.358411 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:32 crc kubenswrapper[5018]: W1014 07:16:32.370730 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode32667c4_6863_453a_a160_9e75fe244463.slice/crio-1345b1cce0ecce6b659950889c68f92a2b7a5875dfe84ba7fbb311d23c9dbb6a WatchSource:0}: Error finding container 1345b1cce0ecce6b659950889c68f92a2b7a5875dfe84ba7fbb311d23c9dbb6a: Status 404 returned error can't find the container with id 1345b1cce0ecce6b659950889c68f92a2b7a5875dfe84ba7fbb311d23c9dbb6a Oct 14 07:16:32 crc kubenswrapper[5018]: I1014 07:16:32.429004 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerStarted","Data":"1345b1cce0ecce6b659950889c68f92a2b7a5875dfe84ba7fbb311d23c9dbb6a"} Oct 14 07:16:33 crc kubenswrapper[5018]: I1014 07:16:33.441571 5018 generic.go:334] "Generic (PLEG): container finished" podID="e32667c4-6863-453a-a160-9e75fe244463" containerID="19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17" exitCode=0 Oct 14 07:16:33 crc kubenswrapper[5018]: I1014 07:16:33.441739 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerDied","Data":"19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17"} Oct 14 07:16:35 crc kubenswrapper[5018]: I1014 07:16:35.468775 5018 generic.go:334] "Generic (PLEG): container finished" podID="e32667c4-6863-453a-a160-9e75fe244463" containerID="f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436" exitCode=0 Oct 14 07:16:35 crc kubenswrapper[5018]: I1014 07:16:35.469023 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerDied","Data":"f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436"} Oct 14 07:16:36 crc kubenswrapper[5018]: I1014 07:16:36.488392 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerStarted","Data":"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01"} Oct 14 07:16:36 crc kubenswrapper[5018]: I1014 07:16:36.520599 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-86hmd" podStartSLOduration=2.886168558 podStartE2EDuration="5.520573556s" podCreationTimestamp="2025-10-14 07:16:31 +0000 UTC" firstStartedPulling="2025-10-14 07:16:33.443572754 +0000 UTC m=+1610.027619421" lastFinishedPulling="2025-10-14 07:16:36.077977792 +0000 UTC m=+1612.662024419" observedRunningTime="2025-10-14 07:16:36.515368608 +0000 UTC m=+1613.099415245" watchObservedRunningTime="2025-10-14 07:16:36.520573556 +0000 UTC m=+1613.104620193" Oct 14 07:16:40 crc kubenswrapper[5018]: I1014 07:16:40.605199 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:16:40 crc kubenswrapper[5018]: E1014 07:16:40.606089 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:16:41 crc kubenswrapper[5018]: I1014 07:16:41.941833 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:41 crc kubenswrapper[5018]: I1014 07:16:41.942692 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:42 crc kubenswrapper[5018]: I1014 07:16:42.022082 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:42 crc kubenswrapper[5018]: I1014 07:16:42.622884 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:42 crc kubenswrapper[5018]: I1014 07:16:42.683073 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:44 crc kubenswrapper[5018]: I1014 07:16:44.560705 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-86hmd" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="registry-server" containerID="cri-o://23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01" gracePeriod=2 Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.019667 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.153268 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities\") pod \"e32667c4-6863-453a-a160-9e75fe244463\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.153339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content\") pod \"e32667c4-6863-453a-a160-9e75fe244463\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.153442 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgxk9\" (UniqueName: \"kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9\") pod \"e32667c4-6863-453a-a160-9e75fe244463\" (UID: \"e32667c4-6863-453a-a160-9e75fe244463\") " Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.155136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities" (OuterVolumeSpecName: "utilities") pod "e32667c4-6863-453a-a160-9e75fe244463" (UID: "e32667c4-6863-453a-a160-9e75fe244463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.165100 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9" (OuterVolumeSpecName: "kube-api-access-jgxk9") pod "e32667c4-6863-453a-a160-9e75fe244463" (UID: "e32667c4-6863-453a-a160-9e75fe244463"). InnerVolumeSpecName "kube-api-access-jgxk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.183036 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e32667c4-6863-453a-a160-9e75fe244463" (UID: "e32667c4-6863-453a-a160-9e75fe244463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.255997 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.256055 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32667c4-6863-453a-a160-9e75fe244463-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.256079 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgxk9\" (UniqueName: \"kubernetes.io/projected/e32667c4-6863-453a-a160-9e75fe244463-kube-api-access-jgxk9\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.573160 5018 generic.go:334] "Generic (PLEG): container finished" podID="e32667c4-6863-453a-a160-9e75fe244463" containerID="23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01" exitCode=0 Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.573219 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86hmd" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.573239 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerDied","Data":"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01"} Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.575039 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86hmd" event={"ID":"e32667c4-6863-453a-a160-9e75fe244463","Type":"ContainerDied","Data":"1345b1cce0ecce6b659950889c68f92a2b7a5875dfe84ba7fbb311d23c9dbb6a"} Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.575100 5018 scope.go:117] "RemoveContainer" containerID="23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.602136 5018 scope.go:117] "RemoveContainer" containerID="f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.628531 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.634953 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-86hmd"] Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.639695 5018 scope.go:117] "RemoveContainer" containerID="19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.660391 5018 scope.go:117] "RemoveContainer" containerID="23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01" Oct 14 07:16:45 crc kubenswrapper[5018]: E1014 07:16:45.660780 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01\": container with ID starting with 23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01 not found: ID does not exist" containerID="23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.660807 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01"} err="failed to get container status \"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01\": rpc error: code = NotFound desc = could not find container \"23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01\": container with ID starting with 23b3b4a8e46f851e17b2b657192778248645c6dc7c9b3bc63f711226332d2c01 not found: ID does not exist" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.660827 5018 scope.go:117] "RemoveContainer" containerID="f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436" Oct 14 07:16:45 crc kubenswrapper[5018]: E1014 07:16:45.661114 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436\": container with ID starting with f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436 not found: ID does not exist" containerID="f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.661136 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436"} err="failed to get container status \"f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436\": rpc error: code = NotFound desc = could not find container \"f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436\": container with ID starting with f714662560d71534517bd7c81de8c1277ed24154f8fe5149713a51dde36f1436 not found: ID does not exist" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.661151 5018 scope.go:117] "RemoveContainer" containerID="19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17" Oct 14 07:16:45 crc kubenswrapper[5018]: E1014 07:16:45.661554 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17\": container with ID starting with 19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17 not found: ID does not exist" containerID="19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17" Oct 14 07:16:45 crc kubenswrapper[5018]: I1014 07:16:45.661607 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17"} err="failed to get container status \"19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17\": rpc error: code = NotFound desc = could not find container \"19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17\": container with ID starting with 19d3943dba166474cbb469ec99ff0ceb359312824e663e628470201ad9de3a17 not found: ID does not exist" Oct 14 07:16:46 crc kubenswrapper[5018]: I1014 07:16:46.620730 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e32667c4-6863-453a-a160-9e75fe244463" path="/var/lib/kubelet/pods/e32667c4-6863-453a-a160-9e75fe244463/volumes" Oct 14 07:16:54 crc kubenswrapper[5018]: I1014 07:16:54.610390 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:16:54 crc kubenswrapper[5018]: E1014 07:16:54.611296 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.380323 5018 scope.go:117] "RemoveContainer" containerID="27d9c4ea59e66a6c005bf07155f6c75635522244901f882a986117cbb32462c3" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.433281 5018 scope.go:117] "RemoveContainer" containerID="57ed62c426c1d9d523dff08f06246a7f633c0647bf1efed3794c0c9655b61674" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.475079 5018 scope.go:117] "RemoveContainer" containerID="bf07b46b7cc5e2c9c4c9ce6bf95b433ad0a7c99598d56f7fc91a388368a3f53f" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.494687 5018 scope.go:117] "RemoveContainer" containerID="fe6322df77a7424d475750b812ebb3bacd32eea8420f037360442254cd9fae19" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.524900 5018 scope.go:117] "RemoveContainer" containerID="0b41efaf38effb59e9892333dafa31b2432c64f40e8aa04ffd1d117c8bf57289" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.547768 5018 scope.go:117] "RemoveContainer" containerID="5deaacab0ff8da05b088f6bd2538d21f0764eafa9e3a35c45d78906b2a2b94c7" Oct 14 07:17:03 crc kubenswrapper[5018]: I1014 07:17:03.585749 5018 scope.go:117] "RemoveContainer" containerID="78d244c57eec48f01d053b75b6b15df44756f1a86d158f3abd8d35b740e667d8" Oct 14 07:17:05 crc kubenswrapper[5018]: I1014 07:17:05.604484 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:17:05 crc kubenswrapper[5018]: E1014 07:17:05.606298 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:17:16 crc kubenswrapper[5018]: I1014 07:17:16.605337 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:17:16 crc kubenswrapper[5018]: E1014 07:17:16.607123 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:17:27 crc kubenswrapper[5018]: I1014 07:17:27.604493 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:17:27 crc kubenswrapper[5018]: E1014 07:17:27.605132 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:17:40 crc kubenswrapper[5018]: I1014 07:17:40.605001 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:17:40 crc kubenswrapper[5018]: E1014 07:17:40.605926 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:17:53 crc kubenswrapper[5018]: I1014 07:17:53.606111 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:17:53 crc kubenswrapper[5018]: E1014 07:17:53.607957 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:18:03 crc kubenswrapper[5018]: I1014 07:18:03.679363 5018 scope.go:117] "RemoveContainer" containerID="116ecd9653cec8dd9db0afce5cfcb53c0fbe1151c252757153b99e432a73a32f" Oct 14 07:18:05 crc kubenswrapper[5018]: I1014 07:18:05.604581 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:18:05 crc kubenswrapper[5018]: E1014 07:18:05.605349 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:18:18 crc kubenswrapper[5018]: I1014 07:18:18.606221 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:18:18 crc kubenswrapper[5018]: E1014 07:18:18.607478 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:18:33 crc kubenswrapper[5018]: I1014 07:18:33.606042 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:18:33 crc kubenswrapper[5018]: E1014 07:18:33.606829 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:18:46 crc kubenswrapper[5018]: I1014 07:18:46.605466 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:18:46 crc kubenswrapper[5018]: E1014 07:18:46.606234 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:18:59 crc kubenswrapper[5018]: I1014 07:18:59.605478 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:18:59 crc kubenswrapper[5018]: E1014 07:18:59.606448 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:19:03 crc kubenswrapper[5018]: I1014 07:19:03.762822 5018 scope.go:117] "RemoveContainer" containerID="dc592a10f1f08b665bdc5af98cb705b332dad979193cef3ba26aeda7526045bf" Oct 14 07:19:03 crc kubenswrapper[5018]: I1014 07:19:03.799179 5018 scope.go:117] "RemoveContainer" containerID="bf2bdb19722ae792c1618512a2809aa12cbbb6af222d67173a5b54630b376efa" Oct 14 07:19:13 crc kubenswrapper[5018]: I1014 07:19:13.605319 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:19:13 crc kubenswrapper[5018]: E1014 07:19:13.606499 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:19:24 crc kubenswrapper[5018]: I1014 07:19:24.614054 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:19:24 crc kubenswrapper[5018]: E1014 07:19:24.614991 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:19:39 crc kubenswrapper[5018]: I1014 07:19:39.605403 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:19:39 crc kubenswrapper[5018]: E1014 07:19:39.606742 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:19:51 crc kubenswrapper[5018]: I1014 07:19:51.605459 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:19:51 crc kubenswrapper[5018]: E1014 07:19:51.606692 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:20:03 crc kubenswrapper[5018]: I1014 07:20:03.605416 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:20:03 crc kubenswrapper[5018]: E1014 07:20:03.606731 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:20:18 crc kubenswrapper[5018]: I1014 07:20:18.604821 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:20:18 crc kubenswrapper[5018]: E1014 07:20:18.605656 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:20:31 crc kubenswrapper[5018]: I1014 07:20:31.605243 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:20:31 crc kubenswrapper[5018]: E1014 07:20:31.606029 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:20:45 crc kubenswrapper[5018]: I1014 07:20:45.606194 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:20:46 crc kubenswrapper[5018]: I1014 07:20:46.804041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95"} Oct 14 07:23:02 crc kubenswrapper[5018]: I1014 07:23:02.463756 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:23:02 crc kubenswrapper[5018]: I1014 07:23:02.464536 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:23:32 crc kubenswrapper[5018]: I1014 07:23:32.463014 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:23:32 crc kubenswrapper[5018]: I1014 07:23:32.463594 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:24:02 crc kubenswrapper[5018]: I1014 07:24:02.463177 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:24:02 crc kubenswrapper[5018]: I1014 07:24:02.463896 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:24:02 crc kubenswrapper[5018]: I1014 07:24:02.463986 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:24:02 crc kubenswrapper[5018]: I1014 07:24:02.464917 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:24:02 crc kubenswrapper[5018]: I1014 07:24:02.465100 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95" gracePeriod=600 Oct 14 07:24:03 crc kubenswrapper[5018]: I1014 07:24:03.590380 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95" exitCode=0 Oct 14 07:24:03 crc kubenswrapper[5018]: I1014 07:24:03.590472 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95"} Oct 14 07:24:03 crc kubenswrapper[5018]: I1014 07:24:03.591136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b"} Oct 14 07:24:03 crc kubenswrapper[5018]: I1014 07:24:03.591165 5018 scope.go:117] "RemoveContainer" containerID="03f22549829b29f749fe18667e6388d92cab54eb082dca47b182c75fa0c67cfb" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.059300 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:22 crc kubenswrapper[5018]: E1014 07:24:22.060239 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="extract-content" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.060257 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="extract-content" Oct 14 07:24:22 crc kubenswrapper[5018]: E1014 07:24:22.060278 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="extract-utilities" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.060287 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="extract-utilities" Oct 14 07:24:22 crc kubenswrapper[5018]: E1014 07:24:22.060314 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="registry-server" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.060323 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="registry-server" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.060478 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32667c4-6863-453a-a160-9e75fe244463" containerName="registry-server" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.061729 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.102293 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.179255 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.179706 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.179744 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmqs\" (UniqueName: \"kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.280934 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.281001 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmqs\" (UniqueName: \"kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.281093 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.281748 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.282018 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.301019 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmqs\" (UniqueName: \"kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs\") pod \"certified-operators-xlwlv\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.387726 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:22 crc kubenswrapper[5018]: I1014 07:24:22.863950 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:23 crc kubenswrapper[5018]: I1014 07:24:23.780341 5018 generic.go:334] "Generic (PLEG): container finished" podID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerID="afd0910873c3bad03e5a19d9f772126f7e2aa8a59c32f0f0a6c83888a0b67692" exitCode=0 Oct 14 07:24:23 crc kubenswrapper[5018]: I1014 07:24:23.780406 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerDied","Data":"afd0910873c3bad03e5a19d9f772126f7e2aa8a59c32f0f0a6c83888a0b67692"} Oct 14 07:24:23 crc kubenswrapper[5018]: I1014 07:24:23.780448 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerStarted","Data":"de3456a1dd5fdefc50b6ca91c63afd1d488b2ec389f091b608a5b16f036ff3a7"} Oct 14 07:24:23 crc kubenswrapper[5018]: I1014 07:24:23.784867 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:24:24 crc kubenswrapper[5018]: I1014 07:24:24.793508 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerStarted","Data":"558af8e26f16d1e35217a99c8f9cb0bfe9a8a4e57c98a9ccd2026d9c3945c25b"} Oct 14 07:24:25 crc kubenswrapper[5018]: I1014 07:24:25.811729 5018 generic.go:334] "Generic (PLEG): container finished" podID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerID="558af8e26f16d1e35217a99c8f9cb0bfe9a8a4e57c98a9ccd2026d9c3945c25b" exitCode=0 Oct 14 07:24:25 crc kubenswrapper[5018]: I1014 07:24:25.811814 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerDied","Data":"558af8e26f16d1e35217a99c8f9cb0bfe9a8a4e57c98a9ccd2026d9c3945c25b"} Oct 14 07:24:26 crc kubenswrapper[5018]: I1014 07:24:26.840410 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerStarted","Data":"6b133759e12320325d47f3a3bb2cdd3bee807eb4d4aaa883a1922eedf474b7bb"} Oct 14 07:24:26 crc kubenswrapper[5018]: I1014 07:24:26.872038 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xlwlv" podStartSLOduration=2.346328023 podStartE2EDuration="4.872012614s" podCreationTimestamp="2025-10-14 07:24:22 +0000 UTC" firstStartedPulling="2025-10-14 07:24:23.784443109 +0000 UTC m=+2080.368489736" lastFinishedPulling="2025-10-14 07:24:26.31012767 +0000 UTC m=+2082.894174327" observedRunningTime="2025-10-14 07:24:26.870128261 +0000 UTC m=+2083.454174938" watchObservedRunningTime="2025-10-14 07:24:26.872012614 +0000 UTC m=+2083.456059241" Oct 14 07:24:32 crc kubenswrapper[5018]: I1014 07:24:32.388068 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:32 crc kubenswrapper[5018]: I1014 07:24:32.388537 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:32 crc kubenswrapper[5018]: I1014 07:24:32.452591 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:32 crc kubenswrapper[5018]: I1014 07:24:32.930846 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.312552 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.316212 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.329677 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.399047 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf8hn\" (UniqueName: \"kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.399118 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.399190 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.500237 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.500316 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf8hn\" (UniqueName: \"kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.500699 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.500744 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.501105 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.501435 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.501870 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xlwlv" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="registry-server" containerID="cri-o://6b133759e12320325d47f3a3bb2cdd3bee807eb4d4aaa883a1922eedf474b7bb" gracePeriod=2 Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.533095 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf8hn\" (UniqueName: \"kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn\") pod \"community-operators-8lnps\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.668655 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.931118 5018 generic.go:334] "Generic (PLEG): container finished" podID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerID="6b133759e12320325d47f3a3bb2cdd3bee807eb4d4aaa883a1922eedf474b7bb" exitCode=0 Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.931154 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerDied","Data":"6b133759e12320325d47f3a3bb2cdd3bee807eb4d4aaa883a1922eedf474b7bb"} Oct 14 07:24:35 crc kubenswrapper[5018]: I1014 07:24:35.963722 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.012016 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content\") pod \"e7803d7b-e54c-493d-9e11-5b794ad0b071\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.012081 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities\") pod \"e7803d7b-e54c-493d-9e11-5b794ad0b071\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.012110 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgmqs\" (UniqueName: \"kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs\") pod \"e7803d7b-e54c-493d-9e11-5b794ad0b071\" (UID: \"e7803d7b-e54c-493d-9e11-5b794ad0b071\") " Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.013724 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities" (OuterVolumeSpecName: "utilities") pod "e7803d7b-e54c-493d-9e11-5b794ad0b071" (UID: "e7803d7b-e54c-493d-9e11-5b794ad0b071"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.017761 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs" (OuterVolumeSpecName: "kube-api-access-tgmqs") pod "e7803d7b-e54c-493d-9e11-5b794ad0b071" (UID: "e7803d7b-e54c-493d-9e11-5b794ad0b071"). InnerVolumeSpecName "kube-api-access-tgmqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.066856 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7803d7b-e54c-493d-9e11-5b794ad0b071" (UID: "e7803d7b-e54c-493d-9e11-5b794ad0b071"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.114708 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.114785 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7803d7b-e54c-493d-9e11-5b794ad0b071-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.114800 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgmqs\" (UniqueName: \"kubernetes.io/projected/e7803d7b-e54c-493d-9e11-5b794ad0b071-kube-api-access-tgmqs\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.333242 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:36 crc kubenswrapper[5018]: W1014 07:24:36.339413 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd17e23a_6fb8_4d62_ab98_54c25526ca0c.slice/crio-b818aaf32004ff8c46679df854a3f0f721d455d6eb2e29a98ff3169884090046 WatchSource:0}: Error finding container b818aaf32004ff8c46679df854a3f0f721d455d6eb2e29a98ff3169884090046: Status 404 returned error can't find the container with id b818aaf32004ff8c46679df854a3f0f721d455d6eb2e29a98ff3169884090046 Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.942673 5018 generic.go:334] "Generic (PLEG): container finished" podID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerID="487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c" exitCode=0 Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.942809 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerDied","Data":"487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c"} Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.943220 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerStarted","Data":"b818aaf32004ff8c46679df854a3f0f721d455d6eb2e29a98ff3169884090046"} Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.948212 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlwlv" event={"ID":"e7803d7b-e54c-493d-9e11-5b794ad0b071","Type":"ContainerDied","Data":"de3456a1dd5fdefc50b6ca91c63afd1d488b2ec389f091b608a5b16f036ff3a7"} Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.948291 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlwlv" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.948303 5018 scope.go:117] "RemoveContainer" containerID="6b133759e12320325d47f3a3bb2cdd3bee807eb4d4aaa883a1922eedf474b7bb" Oct 14 07:24:36 crc kubenswrapper[5018]: I1014 07:24:36.985008 5018 scope.go:117] "RemoveContainer" containerID="558af8e26f16d1e35217a99c8f9cb0bfe9a8a4e57c98a9ccd2026d9c3945c25b" Oct 14 07:24:37 crc kubenswrapper[5018]: I1014 07:24:37.006565 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:37 crc kubenswrapper[5018]: I1014 07:24:37.024371 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xlwlv"] Oct 14 07:24:37 crc kubenswrapper[5018]: I1014 07:24:37.026174 5018 scope.go:117] "RemoveContainer" containerID="afd0910873c3bad03e5a19d9f772126f7e2aa8a59c32f0f0a6c83888a0b67692" Oct 14 07:24:38 crc kubenswrapper[5018]: I1014 07:24:38.623719 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" path="/var/lib/kubelet/pods/e7803d7b-e54c-493d-9e11-5b794ad0b071/volumes" Oct 14 07:24:38 crc kubenswrapper[5018]: I1014 07:24:38.969031 5018 generic.go:334] "Generic (PLEG): container finished" podID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerID="9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba" exitCode=0 Oct 14 07:24:38 crc kubenswrapper[5018]: I1014 07:24:38.969105 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerDied","Data":"9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba"} Oct 14 07:24:39 crc kubenswrapper[5018]: I1014 07:24:39.981385 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerStarted","Data":"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7"} Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.005339 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8lnps" podStartSLOduration=2.5496496840000002 podStartE2EDuration="5.005314762s" podCreationTimestamp="2025-10-14 07:24:35 +0000 UTC" firstStartedPulling="2025-10-14 07:24:36.946565846 +0000 UTC m=+2093.530612513" lastFinishedPulling="2025-10-14 07:24:39.402230934 +0000 UTC m=+2095.986277591" observedRunningTime="2025-10-14 07:24:40.002739639 +0000 UTC m=+2096.586786306" watchObservedRunningTime="2025-10-14 07:24:40.005314762 +0000 UTC m=+2096.589361429" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.310591 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:40 crc kubenswrapper[5018]: E1014 07:24:40.311151 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="extract-content" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.311185 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="extract-content" Oct 14 07:24:40 crc kubenswrapper[5018]: E1014 07:24:40.311209 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="extract-utilities" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.311223 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="extract-utilities" Oct 14 07:24:40 crc kubenswrapper[5018]: E1014 07:24:40.311242 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="registry-server" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.311256 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="registry-server" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.311519 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7803d7b-e54c-493d-9e11-5b794ad0b071" containerName="registry-server" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.313344 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.346566 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.390413 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.390494 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.390553 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmq57\" (UniqueName: \"kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.491361 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmq57\" (UniqueName: \"kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.491434 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.491483 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.491979 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.492065 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.513330 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmq57\" (UniqueName: \"kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57\") pod \"redhat-operators-hcnp9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:40 crc kubenswrapper[5018]: I1014 07:24:40.673501 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:41 crc kubenswrapper[5018]: I1014 07:24:41.090140 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:41 crc kubenswrapper[5018]: W1014 07:24:41.095921 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod970d0068_092b_445f_b84d_c846d2291ec9.slice/crio-163861dd2873f4a83c2fda20751382c3689952747271e9e4c12c033d21d918dd WatchSource:0}: Error finding container 163861dd2873f4a83c2fda20751382c3689952747271e9e4c12c033d21d918dd: Status 404 returned error can't find the container with id 163861dd2873f4a83c2fda20751382c3689952747271e9e4c12c033d21d918dd Oct 14 07:24:41 crc kubenswrapper[5018]: I1014 07:24:41.996807 5018 generic.go:334] "Generic (PLEG): container finished" podID="970d0068-092b-445f-b84d-c846d2291ec9" containerID="fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59" exitCode=0 Oct 14 07:24:41 crc kubenswrapper[5018]: I1014 07:24:41.996870 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerDied","Data":"fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59"} Oct 14 07:24:41 crc kubenswrapper[5018]: I1014 07:24:41.997204 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerStarted","Data":"163861dd2873f4a83c2fda20751382c3689952747271e9e4c12c033d21d918dd"} Oct 14 07:24:43 crc kubenswrapper[5018]: I1014 07:24:43.010348 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerStarted","Data":"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a"} Oct 14 07:24:44 crc kubenswrapper[5018]: I1014 07:24:44.028072 5018 generic.go:334] "Generic (PLEG): container finished" podID="970d0068-092b-445f-b84d-c846d2291ec9" containerID="3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a" exitCode=0 Oct 14 07:24:44 crc kubenswrapper[5018]: I1014 07:24:44.028152 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerDied","Data":"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a"} Oct 14 07:24:45 crc kubenswrapper[5018]: I1014 07:24:45.041443 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerStarted","Data":"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227"} Oct 14 07:24:45 crc kubenswrapper[5018]: I1014 07:24:45.065997 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hcnp9" podStartSLOduration=2.599797362 podStartE2EDuration="5.065981169s" podCreationTimestamp="2025-10-14 07:24:40 +0000 UTC" firstStartedPulling="2025-10-14 07:24:41.998924126 +0000 UTC m=+2098.582970763" lastFinishedPulling="2025-10-14 07:24:44.465107913 +0000 UTC m=+2101.049154570" observedRunningTime="2025-10-14 07:24:45.064058554 +0000 UTC m=+2101.648105191" watchObservedRunningTime="2025-10-14 07:24:45.065981169 +0000 UTC m=+2101.650027806" Oct 14 07:24:45 crc kubenswrapper[5018]: I1014 07:24:45.669516 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:45 crc kubenswrapper[5018]: I1014 07:24:45.669593 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:45 crc kubenswrapper[5018]: I1014 07:24:45.714289 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:46 crc kubenswrapper[5018]: I1014 07:24:46.102102 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.499544 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.500152 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8lnps" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="registry-server" containerID="cri-o://a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7" gracePeriod=2 Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.673735 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.674223 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.758342 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:50 crc kubenswrapper[5018]: I1014 07:24:50.994852 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.079102 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf8hn\" (UniqueName: \"kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn\") pod \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.079166 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities\") pod \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.079207 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content\") pod \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\" (UID: \"fd17e23a-6fb8-4d62-ab98-54c25526ca0c\") " Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.086668 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities" (OuterVolumeSpecName: "utilities") pod "fd17e23a-6fb8-4d62-ab98-54c25526ca0c" (UID: "fd17e23a-6fb8-4d62-ab98-54c25526ca0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.090873 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn" (OuterVolumeSpecName: "kube-api-access-mf8hn") pod "fd17e23a-6fb8-4d62-ab98-54c25526ca0c" (UID: "fd17e23a-6fb8-4d62-ab98-54c25526ca0c"). InnerVolumeSpecName "kube-api-access-mf8hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.102591 5018 generic.go:334] "Generic (PLEG): container finished" podID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerID="a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7" exitCode=0 Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.102694 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lnps" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.102711 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerDied","Data":"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7"} Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.102964 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lnps" event={"ID":"fd17e23a-6fb8-4d62-ab98-54c25526ca0c","Type":"ContainerDied","Data":"b818aaf32004ff8c46679df854a3f0f721d455d6eb2e29a98ff3169884090046"} Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.103010 5018 scope.go:117] "RemoveContainer" containerID="a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.129957 5018 scope.go:117] "RemoveContainer" containerID="9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.144907 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd17e23a-6fb8-4d62-ab98-54c25526ca0c" (UID: "fd17e23a-6fb8-4d62-ab98-54c25526ca0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.158951 5018 scope.go:117] "RemoveContainer" containerID="487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.166921 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.180244 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf8hn\" (UniqueName: \"kubernetes.io/projected/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-kube-api-access-mf8hn\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.180307 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.180316 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd17e23a-6fb8-4d62-ab98-54c25526ca0c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.193587 5018 scope.go:117] "RemoveContainer" containerID="a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7" Oct 14 07:24:51 crc kubenswrapper[5018]: E1014 07:24:51.194084 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7\": container with ID starting with a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7 not found: ID does not exist" containerID="a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.194136 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7"} err="failed to get container status \"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7\": rpc error: code = NotFound desc = could not find container \"a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7\": container with ID starting with a8a572e1060f2fdc9efac1a043922f1e741e31261b3abd364c154292190b1aa7 not found: ID does not exist" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.194167 5018 scope.go:117] "RemoveContainer" containerID="9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba" Oct 14 07:24:51 crc kubenswrapper[5018]: E1014 07:24:51.194523 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba\": container with ID starting with 9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba not found: ID does not exist" containerID="9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.194569 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba"} err="failed to get container status \"9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba\": rpc error: code = NotFound desc = could not find container \"9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba\": container with ID starting with 9285da76ea219def1b7d5d1b75128fbebb92c46871ed2eb1c939fb82c5c8aaba not found: ID does not exist" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.194594 5018 scope.go:117] "RemoveContainer" containerID="487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c" Oct 14 07:24:51 crc kubenswrapper[5018]: E1014 07:24:51.194957 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c\": container with ID starting with 487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c not found: ID does not exist" containerID="487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.194989 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c"} err="failed to get container status \"487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c\": rpc error: code = NotFound desc = could not find container \"487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c\": container with ID starting with 487e45ca3d67efc0af0ce432f320d26bd008e7f138c4d24c13962015653c352c not found: ID does not exist" Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.465495 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:51 crc kubenswrapper[5018]: I1014 07:24:51.478349 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8lnps"] Oct 14 07:24:52 crc kubenswrapper[5018]: I1014 07:24:52.500889 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:52 crc kubenswrapper[5018]: I1014 07:24:52.622295 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" path="/var/lib/kubelet/pods/fd17e23a-6fb8-4d62-ab98-54c25526ca0c/volumes" Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.123441 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hcnp9" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="registry-server" containerID="cri-o://37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227" gracePeriod=2 Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.629312 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.723438 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content\") pod \"970d0068-092b-445f-b84d-c846d2291ec9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.723502 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmq57\" (UniqueName: \"kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57\") pod \"970d0068-092b-445f-b84d-c846d2291ec9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.723560 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities\") pod \"970d0068-092b-445f-b84d-c846d2291ec9\" (UID: \"970d0068-092b-445f-b84d-c846d2291ec9\") " Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.724509 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities" (OuterVolumeSpecName: "utilities") pod "970d0068-092b-445f-b84d-c846d2291ec9" (UID: "970d0068-092b-445f-b84d-c846d2291ec9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.727075 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57" (OuterVolumeSpecName: "kube-api-access-nmq57") pod "970d0068-092b-445f-b84d-c846d2291ec9" (UID: "970d0068-092b-445f-b84d-c846d2291ec9"). InnerVolumeSpecName "kube-api-access-nmq57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.824987 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmq57\" (UniqueName: \"kubernetes.io/projected/970d0068-092b-445f-b84d-c846d2291ec9-kube-api-access-nmq57\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:53 crc kubenswrapper[5018]: I1014 07:24:53.825049 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.104026 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "970d0068-092b-445f-b84d-c846d2291ec9" (UID: "970d0068-092b-445f-b84d-c846d2291ec9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.136802 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970d0068-092b-445f-b84d-c846d2291ec9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.140363 5018 generic.go:334] "Generic (PLEG): container finished" podID="970d0068-092b-445f-b84d-c846d2291ec9" containerID="37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227" exitCode=0 Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.140417 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerDied","Data":"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227"} Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.140449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcnp9" event={"ID":"970d0068-092b-445f-b84d-c846d2291ec9","Type":"ContainerDied","Data":"163861dd2873f4a83c2fda20751382c3689952747271e9e4c12c033d21d918dd"} Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.140472 5018 scope.go:117] "RemoveContainer" containerID="37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.140494 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcnp9" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.169445 5018 scope.go:117] "RemoveContainer" containerID="3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.195126 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.203512 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hcnp9"] Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.216777 5018 scope.go:117] "RemoveContainer" containerID="fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.254645 5018 scope.go:117] "RemoveContainer" containerID="37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227" Oct 14 07:24:54 crc kubenswrapper[5018]: E1014 07:24:54.255702 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227\": container with ID starting with 37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227 not found: ID does not exist" containerID="37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.255755 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227"} err="failed to get container status \"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227\": rpc error: code = NotFound desc = could not find container \"37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227\": container with ID starting with 37a4ddba8ba7fd57fead07a581a116f78004ee6fda367e51b93f5b84d4c25227 not found: ID does not exist" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.255789 5018 scope.go:117] "RemoveContainer" containerID="3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a" Oct 14 07:24:54 crc kubenswrapper[5018]: E1014 07:24:54.256193 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a\": container with ID starting with 3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a not found: ID does not exist" containerID="3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.256219 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a"} err="failed to get container status \"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a\": rpc error: code = NotFound desc = could not find container \"3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a\": container with ID starting with 3a558adc4b047b0d08a1e3c058e3927dd57859d63f1b09eddd12a117eaf0a51a not found: ID does not exist" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.256234 5018 scope.go:117] "RemoveContainer" containerID="fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59" Oct 14 07:24:54 crc kubenswrapper[5018]: E1014 07:24:54.256867 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59\": container with ID starting with fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59 not found: ID does not exist" containerID="fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.256921 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59"} err="failed to get container status \"fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59\": rpc error: code = NotFound desc = could not find container \"fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59\": container with ID starting with fa0d492bf87fabd47585e8728816af3024a835c9b7c97fa9c870930e9ae46a59 not found: ID does not exist" Oct 14 07:24:54 crc kubenswrapper[5018]: I1014 07:24:54.624146 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="970d0068-092b-445f-b84d-c846d2291ec9" path="/var/lib/kubelet/pods/970d0068-092b-445f-b84d-c846d2291ec9/volumes" Oct 14 07:26:02 crc kubenswrapper[5018]: I1014 07:26:02.464195 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:26:02 crc kubenswrapper[5018]: I1014 07:26:02.464908 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:26:32 crc kubenswrapper[5018]: I1014 07:26:32.463838 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:26:32 crc kubenswrapper[5018]: I1014 07:26:32.464767 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.461573 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462653 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462669 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462681 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="extract-utilities" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462689 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="extract-utilities" Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462714 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="extract-utilities" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462722 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="extract-utilities" Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462733 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462741 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462764 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="extract-content" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462772 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="extract-content" Oct 14 07:26:48 crc kubenswrapper[5018]: E1014 07:26:48.462784 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="extract-content" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462790 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="extract-content" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462967 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="970d0068-092b-445f-b84d-c846d2291ec9" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.462990 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd17e23a-6fb8-4d62-ab98-54c25526ca0c" containerName="registry-server" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.465178 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.483706 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.617372 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.617447 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58xss\" (UniqueName: \"kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.617484 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.718452 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.718559 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.718741 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58xss\" (UniqueName: \"kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.719008 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.719771 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.744663 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58xss\" (UniqueName: \"kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss\") pod \"redhat-marketplace-q5t78\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:48 crc kubenswrapper[5018]: I1014 07:26:48.826687 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:49 crc kubenswrapper[5018]: I1014 07:26:49.256350 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:26:49 crc kubenswrapper[5018]: I1014 07:26:49.295948 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerStarted","Data":"895cef814a0c2943de6e15b0fbb3ac458d5ca327797e962c2660d4a7c0db7194"} Oct 14 07:26:50 crc kubenswrapper[5018]: I1014 07:26:50.309499 5018 generic.go:334] "Generic (PLEG): container finished" podID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerID="a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d" exitCode=0 Oct 14 07:26:50 crc kubenswrapper[5018]: I1014 07:26:50.309578 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerDied","Data":"a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d"} Oct 14 07:26:51 crc kubenswrapper[5018]: I1014 07:26:51.323748 5018 generic.go:334] "Generic (PLEG): container finished" podID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerID="4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb" exitCode=0 Oct 14 07:26:51 crc kubenswrapper[5018]: I1014 07:26:51.323812 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerDied","Data":"4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb"} Oct 14 07:26:52 crc kubenswrapper[5018]: I1014 07:26:52.336993 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerStarted","Data":"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87"} Oct 14 07:26:52 crc kubenswrapper[5018]: I1014 07:26:52.369209 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q5t78" podStartSLOduration=2.887698818 podStartE2EDuration="4.369181602s" podCreationTimestamp="2025-10-14 07:26:48 +0000 UTC" firstStartedPulling="2025-10-14 07:26:50.312700068 +0000 UTC m=+2226.896746745" lastFinishedPulling="2025-10-14 07:26:51.794182862 +0000 UTC m=+2228.378229529" observedRunningTime="2025-10-14 07:26:52.362831101 +0000 UTC m=+2228.946877728" watchObservedRunningTime="2025-10-14 07:26:52.369181602 +0000 UTC m=+2228.953228259" Oct 14 07:26:58 crc kubenswrapper[5018]: I1014 07:26:58.827159 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:58 crc kubenswrapper[5018]: I1014 07:26:58.828866 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:58 crc kubenswrapper[5018]: I1014 07:26:58.906234 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:59 crc kubenswrapper[5018]: I1014 07:26:59.463782 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:26:59 crc kubenswrapper[5018]: I1014 07:26:59.528711 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.419697 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q5t78" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="registry-server" containerID="cri-o://3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87" gracePeriod=2 Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.895668 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.988798 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58xss\" (UniqueName: \"kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss\") pod \"711ae49a-4b68-4b0e-b9f1-875713ffad23\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.989196 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content\") pod \"711ae49a-4b68-4b0e-b9f1-875713ffad23\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.989230 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities\") pod \"711ae49a-4b68-4b0e-b9f1-875713ffad23\" (UID: \"711ae49a-4b68-4b0e-b9f1-875713ffad23\") " Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.990756 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities" (OuterVolumeSpecName: "utilities") pod "711ae49a-4b68-4b0e-b9f1-875713ffad23" (UID: "711ae49a-4b68-4b0e-b9f1-875713ffad23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:27:01 crc kubenswrapper[5018]: I1014 07:27:01.994347 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss" (OuterVolumeSpecName: "kube-api-access-58xss") pod "711ae49a-4b68-4b0e-b9f1-875713ffad23" (UID: "711ae49a-4b68-4b0e-b9f1-875713ffad23"). InnerVolumeSpecName "kube-api-access-58xss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.007718 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "711ae49a-4b68-4b0e-b9f1-875713ffad23" (UID: "711ae49a-4b68-4b0e-b9f1-875713ffad23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.092378 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58xss\" (UniqueName: \"kubernetes.io/projected/711ae49a-4b68-4b0e-b9f1-875713ffad23-kube-api-access-58xss\") on node \"crc\" DevicePath \"\"" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.092462 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.092481 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711ae49a-4b68-4b0e-b9f1-875713ffad23-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.433808 5018 generic.go:334] "Generic (PLEG): container finished" podID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerID="3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87" exitCode=0 Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.433899 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerDied","Data":"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87"} Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.433980 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5t78" event={"ID":"711ae49a-4b68-4b0e-b9f1-875713ffad23","Type":"ContainerDied","Data":"895cef814a0c2943de6e15b0fbb3ac458d5ca327797e962c2660d4a7c0db7194"} Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.434011 5018 scope.go:117] "RemoveContainer" containerID="3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.433896 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5t78" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.462987 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.463084 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.463157 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.464234 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.464387 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" gracePeriod=600 Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.469709 5018 scope.go:117] "RemoveContainer" containerID="4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.488486 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.502682 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5t78"] Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.514907 5018 scope.go:117] "RemoveContainer" containerID="a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.555783 5018 scope.go:117] "RemoveContainer" containerID="3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87" Oct 14 07:27:02 crc kubenswrapper[5018]: E1014 07:27:02.556309 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87\": container with ID starting with 3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87 not found: ID does not exist" containerID="3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.556354 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87"} err="failed to get container status \"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87\": rpc error: code = NotFound desc = could not find container \"3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87\": container with ID starting with 3073ebd5aa2417d8fd3f5ab493a4df2e61ae6f79a2079064842ef7639f8b3c87 not found: ID does not exist" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.556383 5018 scope.go:117] "RemoveContainer" containerID="4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb" Oct 14 07:27:02 crc kubenswrapper[5018]: E1014 07:27:02.556788 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb\": container with ID starting with 4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb not found: ID does not exist" containerID="4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.556854 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb"} err="failed to get container status \"4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb\": rpc error: code = NotFound desc = could not find container \"4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb\": container with ID starting with 4537dba0e6460404acb8a78b81aeb696d4ae05c521958b865fb2c6bec58fb6fb not found: ID does not exist" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.556895 5018 scope.go:117] "RemoveContainer" containerID="a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d" Oct 14 07:27:02 crc kubenswrapper[5018]: E1014 07:27:02.557340 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d\": container with ID starting with a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d not found: ID does not exist" containerID="a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.557392 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d"} err="failed to get container status \"a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d\": rpc error: code = NotFound desc = could not find container \"a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d\": container with ID starting with a6160c832beb2675679ce658df8ea59c60ff6a441c1cbada5a5c2a6ec1df095d not found: ID does not exist" Oct 14 07:27:02 crc kubenswrapper[5018]: E1014 07:27:02.589583 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:27:02 crc kubenswrapper[5018]: I1014 07:27:02.614450 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" path="/var/lib/kubelet/pods/711ae49a-4b68-4b0e-b9f1-875713ffad23/volumes" Oct 14 07:27:03 crc kubenswrapper[5018]: I1014 07:27:03.451144 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" exitCode=0 Oct 14 07:27:03 crc kubenswrapper[5018]: I1014 07:27:03.451191 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b"} Oct 14 07:27:03 crc kubenswrapper[5018]: I1014 07:27:03.451323 5018 scope.go:117] "RemoveContainer" containerID="9cdf7e41b842d847636aef464866783615a716687fdd0f3d6684bdabd6c23c95" Oct 14 07:27:03 crc kubenswrapper[5018]: I1014 07:27:03.452052 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:27:03 crc kubenswrapper[5018]: E1014 07:27:03.453151 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:27:18 crc kubenswrapper[5018]: I1014 07:27:18.605943 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:27:18 crc kubenswrapper[5018]: E1014 07:27:18.607244 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:27:32 crc kubenswrapper[5018]: I1014 07:27:32.605296 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:27:32 crc kubenswrapper[5018]: E1014 07:27:32.606192 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:27:45 crc kubenswrapper[5018]: I1014 07:27:45.605052 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:27:45 crc kubenswrapper[5018]: E1014 07:27:45.605481 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:28:00 crc kubenswrapper[5018]: I1014 07:28:00.606247 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:28:00 crc kubenswrapper[5018]: E1014 07:28:00.607214 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:28:14 crc kubenswrapper[5018]: I1014 07:28:14.613552 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:28:14 crc kubenswrapper[5018]: E1014 07:28:14.614280 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:28:26 crc kubenswrapper[5018]: I1014 07:28:26.605471 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:28:26 crc kubenswrapper[5018]: E1014 07:28:26.606355 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:28:41 crc kubenswrapper[5018]: I1014 07:28:41.605293 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:28:41 crc kubenswrapper[5018]: E1014 07:28:41.607475 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:28:55 crc kubenswrapper[5018]: I1014 07:28:55.604863 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:28:55 crc kubenswrapper[5018]: E1014 07:28:55.605696 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:29:07 crc kubenswrapper[5018]: I1014 07:29:07.604903 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:29:07 crc kubenswrapper[5018]: E1014 07:29:07.605891 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:29:22 crc kubenswrapper[5018]: I1014 07:29:22.605235 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:29:22 crc kubenswrapper[5018]: E1014 07:29:22.606419 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:29:34 crc kubenswrapper[5018]: I1014 07:29:34.615546 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:29:34 crc kubenswrapper[5018]: E1014 07:29:34.616915 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:29:46 crc kubenswrapper[5018]: I1014 07:29:46.604516 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:29:46 crc kubenswrapper[5018]: E1014 07:29:46.605468 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:29:59 crc kubenswrapper[5018]: I1014 07:29:59.605343 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:29:59 crc kubenswrapper[5018]: E1014 07:29:59.606277 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.171530 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg"] Oct 14 07:30:00 crc kubenswrapper[5018]: E1014 07:30:00.172139 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.172185 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[5018]: E1014 07:30:00.172233 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.172252 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[5018]: E1014 07:30:00.172307 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.172325 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.172557 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="711ae49a-4b68-4b0e-b9f1-875713ffad23" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.173158 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.176142 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.176330 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.187077 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg"] Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.231502 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-447cr\" (UniqueName: \"kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.231576 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.231747 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.333448 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-447cr\" (UniqueName: \"kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.333574 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.333666 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.335397 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.343530 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.366333 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-447cr\" (UniqueName: \"kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr\") pod \"collect-profiles-29340450-7sdsg\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.494781 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:00 crc kubenswrapper[5018]: I1014 07:30:00.972116 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg"] Oct 14 07:30:01 crc kubenswrapper[5018]: I1014 07:30:01.095942 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" event={"ID":"c6e546e7-fc9e-4719-9b9f-336487c6e19f","Type":"ContainerStarted","Data":"32732d3b666d5d4b6d69aa3feeb938579f95ecae680e24bea682e57791f33f9c"} Oct 14 07:30:02 crc kubenswrapper[5018]: I1014 07:30:02.111757 5018 generic.go:334] "Generic (PLEG): container finished" podID="c6e546e7-fc9e-4719-9b9f-336487c6e19f" containerID="b23306eb3f6d47deb83cd501e2659455c356aede3644c92e83583882f777e7ad" exitCode=0 Oct 14 07:30:02 crc kubenswrapper[5018]: I1014 07:30:02.111854 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" event={"ID":"c6e546e7-fc9e-4719-9b9f-336487c6e19f","Type":"ContainerDied","Data":"b23306eb3f6d47deb83cd501e2659455c356aede3644c92e83583882f777e7ad"} Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.494474 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.596983 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume\") pod \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.597054 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume\") pod \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.597090 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-447cr\" (UniqueName: \"kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr\") pod \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\" (UID: \"c6e546e7-fc9e-4719-9b9f-336487c6e19f\") " Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.598689 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume" (OuterVolumeSpecName: "config-volume") pod "c6e546e7-fc9e-4719-9b9f-336487c6e19f" (UID: "c6e546e7-fc9e-4719-9b9f-336487c6e19f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.608858 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c6e546e7-fc9e-4719-9b9f-336487c6e19f" (UID: "c6e546e7-fc9e-4719-9b9f-336487c6e19f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.611690 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr" (OuterVolumeSpecName: "kube-api-access-447cr") pod "c6e546e7-fc9e-4719-9b9f-336487c6e19f" (UID: "c6e546e7-fc9e-4719-9b9f-336487c6e19f"). InnerVolumeSpecName "kube-api-access-447cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.699006 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6e546e7-fc9e-4719-9b9f-336487c6e19f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.699061 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6e546e7-fc9e-4719-9b9f-336487c6e19f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:03 crc kubenswrapper[5018]: I1014 07:30:03.699082 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-447cr\" (UniqueName: \"kubernetes.io/projected/c6e546e7-fc9e-4719-9b9f-336487c6e19f-kube-api-access-447cr\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.132512 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" event={"ID":"c6e546e7-fc9e-4719-9b9f-336487c6e19f","Type":"ContainerDied","Data":"32732d3b666d5d4b6d69aa3feeb938579f95ecae680e24bea682e57791f33f9c"} Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.132549 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32732d3b666d5d4b6d69aa3feeb938579f95ecae680e24bea682e57791f33f9c" Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.133085 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg" Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.594443 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp"] Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.604289 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340405-2ndbp"] Oct 14 07:30:04 crc kubenswrapper[5018]: I1014 07:30:04.616920 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a8f667-d3bf-43eb-b046-291b70da5c13" path="/var/lib/kubelet/pods/b7a8f667-d3bf-43eb-b046-291b70da5c13/volumes" Oct 14 07:30:12 crc kubenswrapper[5018]: I1014 07:30:12.605601 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:30:12 crc kubenswrapper[5018]: E1014 07:30:12.606723 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:30:24 crc kubenswrapper[5018]: I1014 07:30:24.615194 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:30:24 crc kubenswrapper[5018]: E1014 07:30:24.616401 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:30:36 crc kubenswrapper[5018]: I1014 07:30:36.605478 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:30:36 crc kubenswrapper[5018]: E1014 07:30:36.606511 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:30:49 crc kubenswrapper[5018]: I1014 07:30:49.610837 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:30:49 crc kubenswrapper[5018]: E1014 07:30:49.612106 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:31:00 crc kubenswrapper[5018]: I1014 07:31:00.605907 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:31:00 crc kubenswrapper[5018]: E1014 07:31:00.607922 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:31:04 crc kubenswrapper[5018]: I1014 07:31:04.152739 5018 scope.go:117] "RemoveContainer" containerID="24c829c769ae1e7e691b4788880cac2bc276230b80d85df18f092f5b48c7ba7d" Oct 14 07:31:12 crc kubenswrapper[5018]: I1014 07:31:12.605457 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:31:12 crc kubenswrapper[5018]: E1014 07:31:12.606602 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:31:26 crc kubenswrapper[5018]: I1014 07:31:26.605742 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:31:26 crc kubenswrapper[5018]: E1014 07:31:26.606525 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:31:37 crc kubenswrapper[5018]: I1014 07:31:37.605112 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:31:37 crc kubenswrapper[5018]: E1014 07:31:37.606248 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:31:50 crc kubenswrapper[5018]: I1014 07:31:50.606099 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:31:50 crc kubenswrapper[5018]: E1014 07:31:50.607181 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:32:05 crc kubenswrapper[5018]: I1014 07:32:05.605567 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:32:06 crc kubenswrapper[5018]: I1014 07:32:06.362006 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb"} Oct 14 07:34:32 crc kubenswrapper[5018]: I1014 07:34:32.464055 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:34:32 crc kubenswrapper[5018]: I1014 07:34:32.464700 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.002793 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:34:49 crc kubenswrapper[5018]: E1014 07:34:49.003850 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6e546e7-fc9e-4719-9b9f-336487c6e19f" containerName="collect-profiles" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.003873 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6e546e7-fc9e-4719-9b9f-336487c6e19f" containerName="collect-profiles" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.005220 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6e546e7-fc9e-4719-9b9f-336487c6e19f" containerName="collect-profiles" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.007053 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.029056 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.190297 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.190367 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.190416 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f42t\" (UniqueName: \"kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.291846 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.291939 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.292005 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f42t\" (UniqueName: \"kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.292407 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.292905 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.313253 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f42t\" (UniqueName: \"kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t\") pod \"certified-operators-fj7g2\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.352704 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.591007 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.948404 5018 generic.go:334] "Generic (PLEG): container finished" podID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerID="1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca" exitCode=0 Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.948493 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerDied","Data":"1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca"} Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.949464 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerStarted","Data":"086571febde46e2f1e69d74da46e966e5a63b8c02fde0ddee9b776373e39f31f"} Oct 14 07:34:49 crc kubenswrapper[5018]: I1014 07:34:49.951311 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:34:50 crc kubenswrapper[5018]: I1014 07:34:50.962310 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerStarted","Data":"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f"} Oct 14 07:34:51 crc kubenswrapper[5018]: I1014 07:34:51.973792 5018 generic.go:334] "Generic (PLEG): container finished" podID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerID="adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f" exitCode=0 Oct 14 07:34:51 crc kubenswrapper[5018]: I1014 07:34:51.973838 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerDied","Data":"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f"} Oct 14 07:34:53 crc kubenswrapper[5018]: I1014 07:34:53.998168 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerStarted","Data":"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7"} Oct 14 07:34:59 crc kubenswrapper[5018]: I1014 07:34:59.352907 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:59 crc kubenswrapper[5018]: I1014 07:34:59.353352 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:59 crc kubenswrapper[5018]: I1014 07:34:59.416892 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:34:59 crc kubenswrapper[5018]: I1014 07:34:59.440710 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fj7g2" podStartSLOduration=8.614287206 podStartE2EDuration="11.440689584s" podCreationTimestamp="2025-10-14 07:34:48 +0000 UTC" firstStartedPulling="2025-10-14 07:34:49.9508735 +0000 UTC m=+2706.534920167" lastFinishedPulling="2025-10-14 07:34:52.777275858 +0000 UTC m=+2709.361322545" observedRunningTime="2025-10-14 07:34:54.022197495 +0000 UTC m=+2710.606244192" watchObservedRunningTime="2025-10-14 07:34:59.440689584 +0000 UTC m=+2716.024736221" Oct 14 07:35:00 crc kubenswrapper[5018]: I1014 07:35:00.129428 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:35:00 crc kubenswrapper[5018]: I1014 07:35:00.189427 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.083163 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fj7g2" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="registry-server" containerID="cri-o://540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7" gracePeriod=2 Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.463877 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.464317 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.544945 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.710279 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f42t\" (UniqueName: \"kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t\") pod \"06aa8400-e089-48ac-b177-4fa5dcddbe35\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.710388 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content\") pod \"06aa8400-e089-48ac-b177-4fa5dcddbe35\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.710485 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities\") pod \"06aa8400-e089-48ac-b177-4fa5dcddbe35\" (UID: \"06aa8400-e089-48ac-b177-4fa5dcddbe35\") " Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.711384 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities" (OuterVolumeSpecName: "utilities") pod "06aa8400-e089-48ac-b177-4fa5dcddbe35" (UID: "06aa8400-e089-48ac-b177-4fa5dcddbe35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.717828 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t" (OuterVolumeSpecName: "kube-api-access-8f42t") pod "06aa8400-e089-48ac-b177-4fa5dcddbe35" (UID: "06aa8400-e089-48ac-b177-4fa5dcddbe35"). InnerVolumeSpecName "kube-api-access-8f42t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.754760 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06aa8400-e089-48ac-b177-4fa5dcddbe35" (UID: "06aa8400-e089-48ac-b177-4fa5dcddbe35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.812028 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f42t\" (UniqueName: \"kubernetes.io/projected/06aa8400-e089-48ac-b177-4fa5dcddbe35-kube-api-access-8f42t\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.812070 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:02 crc kubenswrapper[5018]: I1014 07:35:02.812084 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06aa8400-e089-48ac-b177-4fa5dcddbe35-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.092245 5018 generic.go:334] "Generic (PLEG): container finished" podID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerID="540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7" exitCode=0 Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.092293 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj7g2" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.092290 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerDied","Data":"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7"} Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.092463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj7g2" event={"ID":"06aa8400-e089-48ac-b177-4fa5dcddbe35","Type":"ContainerDied","Data":"086571febde46e2f1e69d74da46e966e5a63b8c02fde0ddee9b776373e39f31f"} Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.092521 5018 scope.go:117] "RemoveContainer" containerID="540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.113494 5018 scope.go:117] "RemoveContainer" containerID="adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.134386 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.136521 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fj7g2"] Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.144032 5018 scope.go:117] "RemoveContainer" containerID="1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.168149 5018 scope.go:117] "RemoveContainer" containerID="540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7" Oct 14 07:35:03 crc kubenswrapper[5018]: E1014 07:35:03.168745 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7\": container with ID starting with 540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7 not found: ID does not exist" containerID="540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.168792 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7"} err="failed to get container status \"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7\": rpc error: code = NotFound desc = could not find container \"540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7\": container with ID starting with 540c56e18be3070282c2735b8ef4069242b616a4f8da1f97b10d5a246d1cd6a7 not found: ID does not exist" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.168820 5018 scope.go:117] "RemoveContainer" containerID="adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f" Oct 14 07:35:03 crc kubenswrapper[5018]: E1014 07:35:03.169332 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f\": container with ID starting with adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f not found: ID does not exist" containerID="adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.169377 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f"} err="failed to get container status \"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f\": rpc error: code = NotFound desc = could not find container \"adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f\": container with ID starting with adb1debfe3013b83c9e46fde07278af73852cc90ea99691514abd04105a2b20f not found: ID does not exist" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.169407 5018 scope.go:117] "RemoveContainer" containerID="1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca" Oct 14 07:35:03 crc kubenswrapper[5018]: E1014 07:35:03.169825 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca\": container with ID starting with 1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca not found: ID does not exist" containerID="1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca" Oct 14 07:35:03 crc kubenswrapper[5018]: I1014 07:35:03.169876 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca"} err="failed to get container status \"1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca\": rpc error: code = NotFound desc = could not find container \"1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca\": container with ID starting with 1135bda2409a6008d2f23f68d12aab4470e57a95bb104b2f57bc2c234a9b54ca not found: ID does not exist" Oct 14 07:35:04 crc kubenswrapper[5018]: I1014 07:35:04.625397 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" path="/var/lib/kubelet/pods/06aa8400-e089-48ac-b177-4fa5dcddbe35/volumes" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.439753 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:15 crc kubenswrapper[5018]: E1014 07:35:15.441155 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="extract-content" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.441178 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="extract-content" Oct 14 07:35:15 crc kubenswrapper[5018]: E1014 07:35:15.441227 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="registry-server" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.441239 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="registry-server" Oct 14 07:35:15 crc kubenswrapper[5018]: E1014 07:35:15.441255 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="extract-utilities" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.441264 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="extract-utilities" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.441580 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="06aa8400-e089-48ac-b177-4fa5dcddbe35" containerName="registry-server" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.444205 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.457799 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.613951 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8fqc\" (UniqueName: \"kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.614141 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.614408 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.715712 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8fqc\" (UniqueName: \"kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.716138 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.716207 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.716732 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.716799 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.744978 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8fqc\" (UniqueName: \"kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc\") pod \"redhat-operators-rz52f\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:15 crc kubenswrapper[5018]: I1014 07:35:15.773334 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:16 crc kubenswrapper[5018]: I1014 07:35:16.185453 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:16 crc kubenswrapper[5018]: I1014 07:35:16.251766 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerStarted","Data":"06384ebe23b50bbca5574a3fd31bf03db226cd75ea1c12bafd1d345e156e617a"} Oct 14 07:35:17 crc kubenswrapper[5018]: I1014 07:35:17.259953 5018 generic.go:334] "Generic (PLEG): container finished" podID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerID="623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6" exitCode=0 Oct 14 07:35:17 crc kubenswrapper[5018]: I1014 07:35:17.260020 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerDied","Data":"623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6"} Oct 14 07:35:18 crc kubenswrapper[5018]: I1014 07:35:18.268645 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerStarted","Data":"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452"} Oct 14 07:35:19 crc kubenswrapper[5018]: I1014 07:35:19.281870 5018 generic.go:334] "Generic (PLEG): container finished" podID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerID="25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452" exitCode=0 Oct 14 07:35:19 crc kubenswrapper[5018]: I1014 07:35:19.281925 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerDied","Data":"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452"} Oct 14 07:35:20 crc kubenswrapper[5018]: I1014 07:35:20.293365 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerStarted","Data":"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8"} Oct 14 07:35:20 crc kubenswrapper[5018]: I1014 07:35:20.319139 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rz52f" podStartSLOduration=2.763505109 podStartE2EDuration="5.31908108s" podCreationTimestamp="2025-10-14 07:35:15 +0000 UTC" firstStartedPulling="2025-10-14 07:35:17.262705837 +0000 UTC m=+2733.846752464" lastFinishedPulling="2025-10-14 07:35:19.818281778 +0000 UTC m=+2736.402328435" observedRunningTime="2025-10-14 07:35:20.315706424 +0000 UTC m=+2736.899753081" watchObservedRunningTime="2025-10-14 07:35:20.31908108 +0000 UTC m=+2736.903127747" Oct 14 07:35:25 crc kubenswrapper[5018]: I1014 07:35:25.774105 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:25 crc kubenswrapper[5018]: I1014 07:35:25.774579 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:26 crc kubenswrapper[5018]: I1014 07:35:26.833942 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rz52f" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="registry-server" probeResult="failure" output=< Oct 14 07:35:26 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 07:35:26 crc kubenswrapper[5018]: > Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.685068 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nxpmh"] Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.687866 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.704380 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxpmh"] Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.775397 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9l84\" (UniqueName: \"kubernetes.io/projected/d6278096-b891-4ca6-a1e1-37b2c6a4277c-kube-api-access-g9l84\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.775498 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-utilities\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.775540 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-catalog-content\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.876238 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9l84\" (UniqueName: \"kubernetes.io/projected/d6278096-b891-4ca6-a1e1-37b2c6a4277c-kube-api-access-g9l84\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.876535 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-utilities\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.876658 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-catalog-content\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.877017 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-utilities\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.877104 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6278096-b891-4ca6-a1e1-37b2c6a4277c-catalog-content\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:31 crc kubenswrapper[5018]: I1014 07:35:31.896679 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9l84\" (UniqueName: \"kubernetes.io/projected/d6278096-b891-4ca6-a1e1-37b2c6a4277c-kube-api-access-g9l84\") pod \"community-operators-nxpmh\" (UID: \"d6278096-b891-4ca6-a1e1-37b2c6a4277c\") " pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.019696 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.463532 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.463817 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.464129 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.464654 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.464716 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb" gracePeriod=600 Oct 14 07:35:32 crc kubenswrapper[5018]: I1014 07:35:32.527543 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxpmh"] Oct 14 07:35:32 crc kubenswrapper[5018]: W1014 07:35:32.533330 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6278096_b891_4ca6_a1e1_37b2c6a4277c.slice/crio-146cdac9f224ae4ce50113c7f4d28bb9c3dc82ef01dc3c2770898ca123bd9ab1 WatchSource:0}: Error finding container 146cdac9f224ae4ce50113c7f4d28bb9c3dc82ef01dc3c2770898ca123bd9ab1: Status 404 returned error can't find the container with id 146cdac9f224ae4ce50113c7f4d28bb9c3dc82ef01dc3c2770898ca123bd9ab1 Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.408429 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb" exitCode=0 Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.408506 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb"} Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.408940 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212"} Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.408963 5018 scope.go:117] "RemoveContainer" containerID="32ff87a895a9e5d4f7ad537ee8d1a823aef9f75c52dabbf3e979c78b6bc2661b" Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.411795 5018 generic.go:334] "Generic (PLEG): container finished" podID="d6278096-b891-4ca6-a1e1-37b2c6a4277c" containerID="67ef4c94eb5deec2f1bd3c578a6a2d7b1205d3838f237cd78c81a476a1b5a1d8" exitCode=0 Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.412041 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxpmh" event={"ID":"d6278096-b891-4ca6-a1e1-37b2c6a4277c","Type":"ContainerDied","Data":"67ef4c94eb5deec2f1bd3c578a6a2d7b1205d3838f237cd78c81a476a1b5a1d8"} Oct 14 07:35:33 crc kubenswrapper[5018]: I1014 07:35:33.412069 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxpmh" event={"ID":"d6278096-b891-4ca6-a1e1-37b2c6a4277c","Type":"ContainerStarted","Data":"146cdac9f224ae4ce50113c7f4d28bb9c3dc82ef01dc3c2770898ca123bd9ab1"} Oct 14 07:35:35 crc kubenswrapper[5018]: I1014 07:35:35.823708 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:35 crc kubenswrapper[5018]: I1014 07:35:35.871347 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:37 crc kubenswrapper[5018]: I1014 07:35:37.055474 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:37 crc kubenswrapper[5018]: I1014 07:35:37.457955 5018 generic.go:334] "Generic (PLEG): container finished" podID="d6278096-b891-4ca6-a1e1-37b2c6a4277c" containerID="fb031a4d1d8794de893320f0369d12dac8119b316eda96143f17b2bccd40c03c" exitCode=0 Oct 14 07:35:37 crc kubenswrapper[5018]: I1014 07:35:37.458141 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxpmh" event={"ID":"d6278096-b891-4ca6-a1e1-37b2c6a4277c","Type":"ContainerDied","Data":"fb031a4d1d8794de893320f0369d12dac8119b316eda96143f17b2bccd40c03c"} Oct 14 07:35:37 crc kubenswrapper[5018]: I1014 07:35:37.458473 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rz52f" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="registry-server" containerID="cri-o://788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8" gracePeriod=2 Oct 14 07:35:37 crc kubenswrapper[5018]: I1014 07:35:37.961300 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.074031 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content\") pod \"4712613b-9bf9-46d8-9db3-ddc5d8801463\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.074081 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities\") pod \"4712613b-9bf9-46d8-9db3-ddc5d8801463\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.074119 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8fqc\" (UniqueName: \"kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc\") pod \"4712613b-9bf9-46d8-9db3-ddc5d8801463\" (UID: \"4712613b-9bf9-46d8-9db3-ddc5d8801463\") " Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.075580 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities" (OuterVolumeSpecName: "utilities") pod "4712613b-9bf9-46d8-9db3-ddc5d8801463" (UID: "4712613b-9bf9-46d8-9db3-ddc5d8801463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.082464 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc" (OuterVolumeSpecName: "kube-api-access-p8fqc") pod "4712613b-9bf9-46d8-9db3-ddc5d8801463" (UID: "4712613b-9bf9-46d8-9db3-ddc5d8801463"). InnerVolumeSpecName "kube-api-access-p8fqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.177088 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.177153 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8fqc\" (UniqueName: \"kubernetes.io/projected/4712613b-9bf9-46d8-9db3-ddc5d8801463-kube-api-access-p8fqc\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.185422 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4712613b-9bf9-46d8-9db3-ddc5d8801463" (UID: "4712613b-9bf9-46d8-9db3-ddc5d8801463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.279159 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4712613b-9bf9-46d8-9db3-ddc5d8801463-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.472591 5018 generic.go:334] "Generic (PLEG): container finished" podID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerID="788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8" exitCode=0 Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.472699 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz52f" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.472808 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerDied","Data":"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8"} Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.472926 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz52f" event={"ID":"4712613b-9bf9-46d8-9db3-ddc5d8801463","Type":"ContainerDied","Data":"06384ebe23b50bbca5574a3fd31bf03db226cd75ea1c12bafd1d345e156e617a"} Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.472953 5018 scope.go:117] "RemoveContainer" containerID="788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.477893 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxpmh" event={"ID":"d6278096-b891-4ca6-a1e1-37b2c6a4277c","Type":"ContainerStarted","Data":"fb9d60aa9680b9e4fcb609a2740de40d3693de1c95ceda56f1f70bbc27f0a330"} Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.507596 5018 scope.go:117] "RemoveContainer" containerID="25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.515588 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nxpmh" podStartSLOduration=3.006132435 podStartE2EDuration="7.515546773s" podCreationTimestamp="2025-10-14 07:35:31 +0000 UTC" firstStartedPulling="2025-10-14 07:35:33.413772492 +0000 UTC m=+2749.997819139" lastFinishedPulling="2025-10-14 07:35:37.9231868 +0000 UTC m=+2754.507233477" observedRunningTime="2025-10-14 07:35:38.507273248 +0000 UTC m=+2755.091319875" watchObservedRunningTime="2025-10-14 07:35:38.515546773 +0000 UTC m=+2755.099593440" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.541318 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.545453 5018 scope.go:117] "RemoveContainer" containerID="623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.549614 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rz52f"] Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.561080 5018 scope.go:117] "RemoveContainer" containerID="788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8" Oct 14 07:35:38 crc kubenswrapper[5018]: E1014 07:35:38.562231 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8\": container with ID starting with 788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8 not found: ID does not exist" containerID="788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.562362 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8"} err="failed to get container status \"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8\": rpc error: code = NotFound desc = could not find container \"788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8\": container with ID starting with 788e6336f9e3e09ab6cf16623909e30fd47c3e0d66bd00810a3b0dae9627eff8 not found: ID does not exist" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.562448 5018 scope.go:117] "RemoveContainer" containerID="25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452" Oct 14 07:35:38 crc kubenswrapper[5018]: E1014 07:35:38.562946 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452\": container with ID starting with 25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452 not found: ID does not exist" containerID="25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.563338 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452"} err="failed to get container status \"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452\": rpc error: code = NotFound desc = could not find container \"25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452\": container with ID starting with 25cf27464f70ab350f6baea8c0e7f12507ad2f39e947d344645254816739e452 not found: ID does not exist" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.563459 5018 scope.go:117] "RemoveContainer" containerID="623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6" Oct 14 07:35:38 crc kubenswrapper[5018]: E1014 07:35:38.563922 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6\": container with ID starting with 623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6 not found: ID does not exist" containerID="623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.564039 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6"} err="failed to get container status \"623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6\": rpc error: code = NotFound desc = could not find container \"623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6\": container with ID starting with 623b92b3af4fd35d1be2ae7ae4b9b1c454afe986ce295bab1013ed4ed9741ae6 not found: ID does not exist" Oct 14 07:35:38 crc kubenswrapper[5018]: I1014 07:35:38.613864 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" path="/var/lib/kubelet/pods/4712613b-9bf9-46d8-9db3-ddc5d8801463/volumes" Oct 14 07:35:42 crc kubenswrapper[5018]: I1014 07:35:42.020296 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:42 crc kubenswrapper[5018]: I1014 07:35:42.021266 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:42 crc kubenswrapper[5018]: I1014 07:35:42.103108 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.098724 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nxpmh" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.208030 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxpmh"] Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.272521 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.272909 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lvlpg" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="registry-server" containerID="cri-o://6f7e5b36772fb725cf1c59b62ec20d30f1e1bd0e3b55ca2af90ddaabd078aa8d" gracePeriod=2 Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.663230 5018 generic.go:334] "Generic (PLEG): container finished" podID="574592ff-a962-4052-a21f-a114aa3869ff" containerID="6f7e5b36772fb725cf1c59b62ec20d30f1e1bd0e3b55ca2af90ddaabd078aa8d" exitCode=0 Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.663293 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerDied","Data":"6f7e5b36772fb725cf1c59b62ec20d30f1e1bd0e3b55ca2af90ddaabd078aa8d"} Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.663607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lvlpg" event={"ID":"574592ff-a962-4052-a21f-a114aa3869ff","Type":"ContainerDied","Data":"6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341"} Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.663639 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e9405dbab3297764dba425bb22103233d89b5f5ba9c6908908d08767f198341" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.666268 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.728092 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities\") pod \"574592ff-a962-4052-a21f-a114aa3869ff\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.728206 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ckcr\" (UniqueName: \"kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr\") pod \"574592ff-a962-4052-a21f-a114aa3869ff\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.728254 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content\") pod \"574592ff-a962-4052-a21f-a114aa3869ff\" (UID: \"574592ff-a962-4052-a21f-a114aa3869ff\") " Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.728763 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities" (OuterVolumeSpecName: "utilities") pod "574592ff-a962-4052-a21f-a114aa3869ff" (UID: "574592ff-a962-4052-a21f-a114aa3869ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.733694 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr" (OuterVolumeSpecName: "kube-api-access-6ckcr") pod "574592ff-a962-4052-a21f-a114aa3869ff" (UID: "574592ff-a962-4052-a21f-a114aa3869ff"). InnerVolumeSpecName "kube-api-access-6ckcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.780017 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "574592ff-a962-4052-a21f-a114aa3869ff" (UID: "574592ff-a962-4052-a21f-a114aa3869ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.829647 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ckcr\" (UniqueName: \"kubernetes.io/projected/574592ff-a962-4052-a21f-a114aa3869ff-kube-api-access-6ckcr\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.829680 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:52 crc kubenswrapper[5018]: I1014 07:35:52.829691 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574592ff-a962-4052-a21f-a114aa3869ff-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:53 crc kubenswrapper[5018]: I1014 07:35:53.671187 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lvlpg" Oct 14 07:35:53 crc kubenswrapper[5018]: I1014 07:35:53.718742 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 07:35:53 crc kubenswrapper[5018]: I1014 07:35:53.724568 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lvlpg"] Oct 14 07:35:54 crc kubenswrapper[5018]: I1014 07:35:54.616307 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574592ff-a962-4052-a21f-a114aa3869ff" path="/var/lib/kubelet/pods/574592ff-a962-4052-a21f-a114aa3869ff/volumes" Oct 14 07:36:04 crc kubenswrapper[5018]: I1014 07:36:04.294149 5018 scope.go:117] "RemoveContainer" containerID="22bd5dc428dee0e158c5723a6542f71c4bc9bb13b14d4921138c22adac5c3224" Oct 14 07:36:04 crc kubenswrapper[5018]: I1014 07:36:04.325852 5018 scope.go:117] "RemoveContainer" containerID="6f7e5b36772fb725cf1c59b62ec20d30f1e1bd0e3b55ca2af90ddaabd078aa8d" Oct 14 07:36:04 crc kubenswrapper[5018]: I1014 07:36:04.373477 5018 scope.go:117] "RemoveContainer" containerID="c253c6fedb9642a2dde1de157bb555c5a0b393195a9aacda178a94c732244b63" Oct 14 07:37:32 crc kubenswrapper[5018]: I1014 07:37:32.463119 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:37:32 crc kubenswrapper[5018]: I1014 07:37:32.463698 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:38:02 crc kubenswrapper[5018]: I1014 07:38:02.463925 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:38:02 crc kubenswrapper[5018]: I1014 07:38:02.464479 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:38:32 crc kubenswrapper[5018]: I1014 07:38:32.464093 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:38:32 crc kubenswrapper[5018]: I1014 07:38:32.464824 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:38:32 crc kubenswrapper[5018]: I1014 07:38:32.464898 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:38:32 crc kubenswrapper[5018]: I1014 07:38:32.466009 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:38:32 crc kubenswrapper[5018]: I1014 07:38:32.466107 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" gracePeriod=600 Oct 14 07:38:32 crc kubenswrapper[5018]: E1014 07:38:32.612051 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:38:33 crc kubenswrapper[5018]: I1014 07:38:33.153702 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" exitCode=0 Oct 14 07:38:33 crc kubenswrapper[5018]: I1014 07:38:33.153787 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212"} Oct 14 07:38:33 crc kubenswrapper[5018]: I1014 07:38:33.153861 5018 scope.go:117] "RemoveContainer" containerID="0ca153b54f502e40203a3df2b878f23221551681f69473894972bb502c335cdb" Oct 14 07:38:33 crc kubenswrapper[5018]: I1014 07:38:33.154465 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:38:33 crc kubenswrapper[5018]: E1014 07:38:33.154941 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:38:44 crc kubenswrapper[5018]: I1014 07:38:44.616149 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:38:44 crc kubenswrapper[5018]: E1014 07:38:44.617360 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:38:55 crc kubenswrapper[5018]: I1014 07:38:55.605271 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:38:55 crc kubenswrapper[5018]: E1014 07:38:55.606273 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:39:08 crc kubenswrapper[5018]: I1014 07:39:08.605588 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:39:08 crc kubenswrapper[5018]: E1014 07:39:08.606773 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:39:20 crc kubenswrapper[5018]: I1014 07:39:20.607051 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:39:20 crc kubenswrapper[5018]: E1014 07:39:20.608229 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:39:32 crc kubenswrapper[5018]: I1014 07:39:32.605199 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:39:32 crc kubenswrapper[5018]: E1014 07:39:32.607738 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:39:44 crc kubenswrapper[5018]: I1014 07:39:44.619153 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:39:44 crc kubenswrapper[5018]: E1014 07:39:44.620224 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:39:56 crc kubenswrapper[5018]: I1014 07:39:56.604891 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:39:56 crc kubenswrapper[5018]: E1014 07:39:56.606096 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:40:11 crc kubenswrapper[5018]: I1014 07:40:11.604935 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:40:11 crc kubenswrapper[5018]: E1014 07:40:11.605784 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:40:22 crc kubenswrapper[5018]: I1014 07:40:22.604826 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:40:22 crc kubenswrapper[5018]: E1014 07:40:22.606059 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:40:36 crc kubenswrapper[5018]: I1014 07:40:36.605797 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:40:36 crc kubenswrapper[5018]: E1014 07:40:36.608660 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.309889 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310665 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="extract-utilities" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310679 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="extract-utilities" Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310698 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310704 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310712 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="extract-utilities" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310717 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="extract-utilities" Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310738 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="extract-content" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310744 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="extract-content" Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310758 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="extract-content" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310764 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="extract-content" Oct 14 07:40:45 crc kubenswrapper[5018]: E1014 07:40:45.310770 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.310776 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.311027 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4712613b-9bf9-46d8-9db3-ddc5d8801463" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.311041 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="574592ff-a962-4052-a21f-a114aa3869ff" containerName="registry-server" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.313848 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.325585 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.406469 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.406521 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9c5\" (UniqueName: \"kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.406562 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.508424 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.508712 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9c5\" (UniqueName: \"kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.508820 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.509059 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.509365 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.528531 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9c5\" (UniqueName: \"kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5\") pod \"redhat-marketplace-pz5w8\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:45 crc kubenswrapper[5018]: I1014 07:40:45.643878 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:46 crc kubenswrapper[5018]: I1014 07:40:46.143522 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:40:46 crc kubenswrapper[5018]: I1014 07:40:46.431290 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerID="01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35" exitCode=0 Oct 14 07:40:46 crc kubenswrapper[5018]: I1014 07:40:46.431640 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerDied","Data":"01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35"} Oct 14 07:40:46 crc kubenswrapper[5018]: I1014 07:40:46.431786 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerStarted","Data":"99da0fa0d62fd6a0d3bf91f4967cd8e3a6f0bc5b871f33523e7a5e7d623f4876"} Oct 14 07:40:46 crc kubenswrapper[5018]: I1014 07:40:46.433337 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:40:47 crc kubenswrapper[5018]: I1014 07:40:47.605843 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:40:47 crc kubenswrapper[5018]: E1014 07:40:47.606449 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:40:48 crc kubenswrapper[5018]: I1014 07:40:48.452849 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerID="4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac" exitCode=0 Oct 14 07:40:48 crc kubenswrapper[5018]: I1014 07:40:48.452919 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerDied","Data":"4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac"} Oct 14 07:40:49 crc kubenswrapper[5018]: I1014 07:40:49.469672 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerStarted","Data":"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264"} Oct 14 07:40:49 crc kubenswrapper[5018]: I1014 07:40:49.502281 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pz5w8" podStartSLOduration=1.941991413 podStartE2EDuration="4.502250846s" podCreationTimestamp="2025-10-14 07:40:45 +0000 UTC" firstStartedPulling="2025-10-14 07:40:46.432979601 +0000 UTC m=+3063.017026228" lastFinishedPulling="2025-10-14 07:40:48.993239024 +0000 UTC m=+3065.577285661" observedRunningTime="2025-10-14 07:40:49.491590513 +0000 UTC m=+3066.075637180" watchObservedRunningTime="2025-10-14 07:40:49.502250846 +0000 UTC m=+3066.086297523" Oct 14 07:40:55 crc kubenswrapper[5018]: I1014 07:40:55.644231 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:55 crc kubenswrapper[5018]: I1014 07:40:55.644589 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:55 crc kubenswrapper[5018]: I1014 07:40:55.729278 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:40:56 crc kubenswrapper[5018]: I1014 07:40:56.602677 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:41:01 crc kubenswrapper[5018]: I1014 07:41:01.605885 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:41:01 crc kubenswrapper[5018]: E1014 07:41:01.606713 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.082571 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.083651 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pz5w8" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="registry-server" containerID="cri-o://862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264" gracePeriod=2 Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.508763 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.609684 5018 generic.go:334] "Generic (PLEG): container finished" podID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerID="862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264" exitCode=0 Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.609737 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerDied","Data":"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264"} Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.609774 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pz5w8" event={"ID":"6ac0f0db-9d99-40e2-80ae-106f26d4b900","Type":"ContainerDied","Data":"99da0fa0d62fd6a0d3bf91f4967cd8e3a6f0bc5b871f33523e7a5e7d623f4876"} Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.609796 5018 scope.go:117] "RemoveContainer" containerID="862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.609885 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pz5w8" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.618434 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities\") pod \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.618497 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content\") pod \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.618553 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg9c5\" (UniqueName: \"kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5\") pod \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\" (UID: \"6ac0f0db-9d99-40e2-80ae-106f26d4b900\") " Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.619938 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities" (OuterVolumeSpecName: "utilities") pod "6ac0f0db-9d99-40e2-80ae-106f26d4b900" (UID: "6ac0f0db-9d99-40e2-80ae-106f26d4b900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.628794 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5" (OuterVolumeSpecName: "kube-api-access-wg9c5") pod "6ac0f0db-9d99-40e2-80ae-106f26d4b900" (UID: "6ac0f0db-9d99-40e2-80ae-106f26d4b900"). InnerVolumeSpecName "kube-api-access-wg9c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.631267 5018 scope.go:117] "RemoveContainer" containerID="4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.647902 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ac0f0db-9d99-40e2-80ae-106f26d4b900" (UID: "6ac0f0db-9d99-40e2-80ae-106f26d4b900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.664204 5018 scope.go:117] "RemoveContainer" containerID="01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.686588 5018 scope.go:117] "RemoveContainer" containerID="862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264" Oct 14 07:41:03 crc kubenswrapper[5018]: E1014 07:41:03.687036 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264\": container with ID starting with 862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264 not found: ID does not exist" containerID="862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.687065 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264"} err="failed to get container status \"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264\": rpc error: code = NotFound desc = could not find container \"862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264\": container with ID starting with 862191e9054c862b48ab136d7f1602bfb6355bbb1cf423a1439a434a9160f264 not found: ID does not exist" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.687091 5018 scope.go:117] "RemoveContainer" containerID="4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac" Oct 14 07:41:03 crc kubenswrapper[5018]: E1014 07:41:03.687496 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac\": container with ID starting with 4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac not found: ID does not exist" containerID="4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.687537 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac"} err="failed to get container status \"4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac\": rpc error: code = NotFound desc = could not find container \"4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac\": container with ID starting with 4c7cb862c01aff5b07a5376cd9dca48940bb4684a4db0b8b39d8987453d917ac not found: ID does not exist" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.687565 5018 scope.go:117] "RemoveContainer" containerID="01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35" Oct 14 07:41:03 crc kubenswrapper[5018]: E1014 07:41:03.687876 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35\": container with ID starting with 01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35 not found: ID does not exist" containerID="01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.687897 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35"} err="failed to get container status \"01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35\": rpc error: code = NotFound desc = could not find container \"01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35\": container with ID starting with 01ebd65dc4377374a622d76e3823da4a51d25306c99a25cbd5cffde507d77f35 not found: ID does not exist" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.719876 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.719913 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ac0f0db-9d99-40e2-80ae-106f26d4b900-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.719926 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg9c5\" (UniqueName: \"kubernetes.io/projected/6ac0f0db-9d99-40e2-80ae-106f26d4b900-kube-api-access-wg9c5\") on node \"crc\" DevicePath \"\"" Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.976194 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:41:03 crc kubenswrapper[5018]: I1014 07:41:03.976276 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pz5w8"] Oct 14 07:41:04 crc kubenswrapper[5018]: I1014 07:41:04.619238 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" path="/var/lib/kubelet/pods/6ac0f0db-9d99-40e2-80ae-106f26d4b900/volumes" Oct 14 07:41:13 crc kubenswrapper[5018]: I1014 07:41:13.605941 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:41:13 crc kubenswrapper[5018]: E1014 07:41:13.607166 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:41:27 crc kubenswrapper[5018]: I1014 07:41:27.606272 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:41:27 crc kubenswrapper[5018]: E1014 07:41:27.607111 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:41:38 crc kubenswrapper[5018]: I1014 07:41:38.605608 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:41:38 crc kubenswrapper[5018]: E1014 07:41:38.606852 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:41:49 crc kubenswrapper[5018]: I1014 07:41:49.605936 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:41:49 crc kubenswrapper[5018]: E1014 07:41:49.606886 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:42:03 crc kubenswrapper[5018]: I1014 07:42:03.605877 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:42:03 crc kubenswrapper[5018]: E1014 07:42:03.606979 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:42:18 crc kubenswrapper[5018]: I1014 07:42:18.606025 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:42:18 crc kubenswrapper[5018]: E1014 07:42:18.607180 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:42:33 crc kubenswrapper[5018]: I1014 07:42:33.605391 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:42:33 crc kubenswrapper[5018]: E1014 07:42:33.606074 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:42:48 crc kubenswrapper[5018]: I1014 07:42:48.604666 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:42:48 crc kubenswrapper[5018]: E1014 07:42:48.606793 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:43:02 crc kubenswrapper[5018]: I1014 07:43:02.605271 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:43:02 crc kubenswrapper[5018]: E1014 07:43:02.606272 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:43:14 crc kubenswrapper[5018]: I1014 07:43:14.612787 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:43:14 crc kubenswrapper[5018]: E1014 07:43:14.613768 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:43:26 crc kubenswrapper[5018]: I1014 07:43:26.605407 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:43:26 crc kubenswrapper[5018]: E1014 07:43:26.606870 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:43:41 crc kubenswrapper[5018]: I1014 07:43:41.604867 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:43:42 crc kubenswrapper[5018]: I1014 07:43:42.227313 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a"} Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.224792 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk"] Oct 14 07:45:00 crc kubenswrapper[5018]: E1014 07:45:00.225821 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.225839 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[5018]: E1014 07:45:00.225869 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.225877 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[5018]: E1014 07:45:00.225893 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.225902 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.226109 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac0f0db-9d99-40e2-80ae-106f26d4b900" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.226754 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.229932 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.230066 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.237104 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk"] Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.339137 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgvsw\" (UniqueName: \"kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.339443 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.339601 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.440767 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgvsw\" (UniqueName: \"kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.441227 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.441517 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.443020 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.451288 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.458792 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgvsw\" (UniqueName: \"kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw\") pod \"collect-profiles-29340465-mmxdk\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:00 crc kubenswrapper[5018]: I1014 07:45:00.555262 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:01 crc kubenswrapper[5018]: I1014 07:45:01.036325 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk"] Oct 14 07:45:02 crc kubenswrapper[5018]: I1014 07:45:02.007196 5018 generic.go:334] "Generic (PLEG): container finished" podID="8a042892-04be-4f31-8f45-db623910ed9e" containerID="6214bd48018768a8aba58b50dba689a5cdd48223db732fb3a9a50fd68a65d538" exitCode=0 Oct 14 07:45:02 crc kubenswrapper[5018]: I1014 07:45:02.007327 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" event={"ID":"8a042892-04be-4f31-8f45-db623910ed9e","Type":"ContainerDied","Data":"6214bd48018768a8aba58b50dba689a5cdd48223db732fb3a9a50fd68a65d538"} Oct 14 07:45:02 crc kubenswrapper[5018]: I1014 07:45:02.007715 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" event={"ID":"8a042892-04be-4f31-8f45-db623910ed9e","Type":"ContainerStarted","Data":"301760f6f3631306961ee8ab47ceb3af32dab95493a3b0c819438105b6e95099"} Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.350110 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.405531 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume\") pod \"8a042892-04be-4f31-8f45-db623910ed9e\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.405848 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgvsw\" (UniqueName: \"kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw\") pod \"8a042892-04be-4f31-8f45-db623910ed9e\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.405916 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume\") pod \"8a042892-04be-4f31-8f45-db623910ed9e\" (UID: \"8a042892-04be-4f31-8f45-db623910ed9e\") " Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.407061 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume" (OuterVolumeSpecName: "config-volume") pod "8a042892-04be-4f31-8f45-db623910ed9e" (UID: "8a042892-04be-4f31-8f45-db623910ed9e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.414444 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8a042892-04be-4f31-8f45-db623910ed9e" (UID: "8a042892-04be-4f31-8f45-db623910ed9e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.414916 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw" (OuterVolumeSpecName: "kube-api-access-tgvsw") pod "8a042892-04be-4f31-8f45-db623910ed9e" (UID: "8a042892-04be-4f31-8f45-db623910ed9e"). InnerVolumeSpecName "kube-api-access-tgvsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.507867 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgvsw\" (UniqueName: \"kubernetes.io/projected/8a042892-04be-4f31-8f45-db623910ed9e-kube-api-access-tgvsw\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.507899 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a042892-04be-4f31-8f45-db623910ed9e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:03 crc kubenswrapper[5018]: I1014 07:45:03.507908 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a042892-04be-4f31-8f45-db623910ed9e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.029821 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" event={"ID":"8a042892-04be-4f31-8f45-db623910ed9e","Type":"ContainerDied","Data":"301760f6f3631306961ee8ab47ceb3af32dab95493a3b0c819438105b6e95099"} Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.029890 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk" Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.029892 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="301760f6f3631306961ee8ab47ceb3af32dab95493a3b0c819438105b6e95099" Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.472304 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb"] Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.478839 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-kgksb"] Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.614574 5018 scope.go:117] "RemoveContainer" containerID="fc272a2f09dca504610589f45b601b615ee3378941d217d7fe9469fa82e9d158" Oct 14 07:45:04 crc kubenswrapper[5018]: I1014 07:45:04.621502 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc" path="/var/lib/kubelet/pods/4e39fcd9-d0db-41e4-a259-c0cd3e2a01cc/volumes" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.727742 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:05 crc kubenswrapper[5018]: E1014 07:45:05.728353 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a042892-04be-4f31-8f45-db623910ed9e" containerName="collect-profiles" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.728387 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a042892-04be-4f31-8f45-db623910ed9e" containerName="collect-profiles" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.728860 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a042892-04be-4f31-8f45-db623910ed9e" containerName="collect-profiles" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.731275 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.750982 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.844261 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.844373 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.844405 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lwwk\" (UniqueName: \"kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.945415 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.945561 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.946098 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.946138 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lwwk\" (UniqueName: \"kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.946187 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:05 crc kubenswrapper[5018]: I1014 07:45:05.972949 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lwwk\" (UniqueName: \"kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk\") pod \"certified-operators-9zf6k\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:06 crc kubenswrapper[5018]: I1014 07:45:06.062583 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:06 crc kubenswrapper[5018]: I1014 07:45:06.654349 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:07 crc kubenswrapper[5018]: I1014 07:45:07.060727 5018 generic.go:334] "Generic (PLEG): container finished" podID="36ea8c0c-56af-4897-982f-c08c1769694f" containerID="0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a" exitCode=0 Oct 14 07:45:07 crc kubenswrapper[5018]: I1014 07:45:07.060961 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerDied","Data":"0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a"} Oct 14 07:45:07 crc kubenswrapper[5018]: I1014 07:45:07.061142 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerStarted","Data":"162278b5a8e692adb0acd2ff082ffa98cadaa8ff1753d86da6c15786f8a96013"} Oct 14 07:45:08 crc kubenswrapper[5018]: I1014 07:45:08.070453 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerStarted","Data":"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc"} Oct 14 07:45:09 crc kubenswrapper[5018]: I1014 07:45:09.083562 5018 generic.go:334] "Generic (PLEG): container finished" podID="36ea8c0c-56af-4897-982f-c08c1769694f" containerID="00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc" exitCode=0 Oct 14 07:45:09 crc kubenswrapper[5018]: I1014 07:45:09.083612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerDied","Data":"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc"} Oct 14 07:45:10 crc kubenswrapper[5018]: I1014 07:45:10.091878 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerStarted","Data":"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269"} Oct 14 07:45:10 crc kubenswrapper[5018]: I1014 07:45:10.111084 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9zf6k" podStartSLOduration=2.673838697 podStartE2EDuration="5.111064248s" podCreationTimestamp="2025-10-14 07:45:05 +0000 UTC" firstStartedPulling="2025-10-14 07:45:07.063586034 +0000 UTC m=+3323.647632701" lastFinishedPulling="2025-10-14 07:45:09.500811615 +0000 UTC m=+3326.084858252" observedRunningTime="2025-10-14 07:45:10.108017931 +0000 UTC m=+3326.692064558" watchObservedRunningTime="2025-10-14 07:45:10.111064248 +0000 UTC m=+3326.695110875" Oct 14 07:45:16 crc kubenswrapper[5018]: I1014 07:45:16.063739 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:16 crc kubenswrapper[5018]: I1014 07:45:16.064218 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:16 crc kubenswrapper[5018]: I1014 07:45:16.136574 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:16 crc kubenswrapper[5018]: I1014 07:45:16.207655 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:16 crc kubenswrapper[5018]: I1014 07:45:16.385694 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.158867 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9zf6k" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="registry-server" containerID="cri-o://bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269" gracePeriod=2 Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.584684 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.659065 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lwwk\" (UniqueName: \"kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk\") pod \"36ea8c0c-56af-4897-982f-c08c1769694f\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.659193 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities\") pod \"36ea8c0c-56af-4897-982f-c08c1769694f\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.659233 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content\") pod \"36ea8c0c-56af-4897-982f-c08c1769694f\" (UID: \"36ea8c0c-56af-4897-982f-c08c1769694f\") " Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.660326 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities" (OuterVolumeSpecName: "utilities") pod "36ea8c0c-56af-4897-982f-c08c1769694f" (UID: "36ea8c0c-56af-4897-982f-c08c1769694f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.661294 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.665828 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk" (OuterVolumeSpecName: "kube-api-access-9lwwk") pod "36ea8c0c-56af-4897-982f-c08c1769694f" (UID: "36ea8c0c-56af-4897-982f-c08c1769694f"). InnerVolumeSpecName "kube-api-access-9lwwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.762838 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lwwk\" (UniqueName: \"kubernetes.io/projected/36ea8c0c-56af-4897-982f-c08c1769694f-kube-api-access-9lwwk\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.855967 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36ea8c0c-56af-4897-982f-c08c1769694f" (UID: "36ea8c0c-56af-4897-982f-c08c1769694f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:18 crc kubenswrapper[5018]: I1014 07:45:18.864700 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36ea8c0c-56af-4897-982f-c08c1769694f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.173684 5018 generic.go:334] "Generic (PLEG): container finished" podID="36ea8c0c-56af-4897-982f-c08c1769694f" containerID="bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269" exitCode=0 Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.173763 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerDied","Data":"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269"} Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.173813 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zf6k" event={"ID":"36ea8c0c-56af-4897-982f-c08c1769694f","Type":"ContainerDied","Data":"162278b5a8e692adb0acd2ff082ffa98cadaa8ff1753d86da6c15786f8a96013"} Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.173848 5018 scope.go:117] "RemoveContainer" containerID="bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.174954 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zf6k" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.212657 5018 scope.go:117] "RemoveContainer" containerID="00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.233128 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.245012 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9zf6k"] Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.257551 5018 scope.go:117] "RemoveContainer" containerID="0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.299237 5018 scope.go:117] "RemoveContainer" containerID="bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269" Oct 14 07:45:19 crc kubenswrapper[5018]: E1014 07:45:19.299947 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269\": container with ID starting with bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269 not found: ID does not exist" containerID="bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.300000 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269"} err="failed to get container status \"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269\": rpc error: code = NotFound desc = could not find container \"bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269\": container with ID starting with bccefe66eecc87cbfc2680745ef2f8a5431d89cfca0da557743c5cbf47811269 not found: ID does not exist" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.300036 5018 scope.go:117] "RemoveContainer" containerID="00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc" Oct 14 07:45:19 crc kubenswrapper[5018]: E1014 07:45:19.300710 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc\": container with ID starting with 00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc not found: ID does not exist" containerID="00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.300915 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc"} err="failed to get container status \"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc\": rpc error: code = NotFound desc = could not find container \"00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc\": container with ID starting with 00068d2ca15badf3c150939e237828b99283e2c57a1cbc39d03849ddf4aa18fc not found: ID does not exist" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.301083 5018 scope.go:117] "RemoveContainer" containerID="0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a" Oct 14 07:45:19 crc kubenswrapper[5018]: E1014 07:45:19.301663 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a\": container with ID starting with 0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a not found: ID does not exist" containerID="0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a" Oct 14 07:45:19 crc kubenswrapper[5018]: I1014 07:45:19.301708 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a"} err="failed to get container status \"0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a\": rpc error: code = NotFound desc = could not find container \"0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a\": container with ID starting with 0bc4a364d0448111af4929d8075275f3eb94ec2124665f9cf7aa7d31af10da4a not found: ID does not exist" Oct 14 07:45:20 crc kubenswrapper[5018]: I1014 07:45:20.621479 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" path="/var/lib/kubelet/pods/36ea8c0c-56af-4897-982f-c08c1769694f/volumes" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.365429 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:25 crc kubenswrapper[5018]: E1014 07:45:25.366922 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="registry-server" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.366939 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="registry-server" Oct 14 07:45:25 crc kubenswrapper[5018]: E1014 07:45:25.366970 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="extract-content" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.366979 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="extract-content" Oct 14 07:45:25 crc kubenswrapper[5018]: E1014 07:45:25.367001 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="extract-utilities" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.367011 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="extract-utilities" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.367182 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="36ea8c0c-56af-4897-982f-c08c1769694f" containerName="registry-server" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.375591 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.385482 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.474679 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.474756 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4468d\" (UniqueName: \"kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.474996 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.576062 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.576433 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.576463 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4468d\" (UniqueName: \"kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.576899 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.577002 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.598136 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4468d\" (UniqueName: \"kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d\") pod \"redhat-operators-vcpks\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:25 crc kubenswrapper[5018]: I1014 07:45:25.707898 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:26 crc kubenswrapper[5018]: I1014 07:45:26.156830 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:26 crc kubenswrapper[5018]: I1014 07:45:26.250985 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerStarted","Data":"fde75702ae805e72da6f0c0b1bad965017ab75d02195dd67ff8b818ffc16131f"} Oct 14 07:45:27 crc kubenswrapper[5018]: I1014 07:45:27.262528 5018 generic.go:334] "Generic (PLEG): container finished" podID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerID="3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3" exitCode=0 Oct 14 07:45:27 crc kubenswrapper[5018]: I1014 07:45:27.262687 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerDied","Data":"3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3"} Oct 14 07:45:28 crc kubenswrapper[5018]: I1014 07:45:28.275201 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerStarted","Data":"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c"} Oct 14 07:45:29 crc kubenswrapper[5018]: I1014 07:45:29.295929 5018 generic.go:334] "Generic (PLEG): container finished" podID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerID="fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c" exitCode=0 Oct 14 07:45:29 crc kubenswrapper[5018]: I1014 07:45:29.296122 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerDied","Data":"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c"} Oct 14 07:45:30 crc kubenswrapper[5018]: I1014 07:45:30.307406 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerStarted","Data":"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f"} Oct 14 07:45:30 crc kubenswrapper[5018]: I1014 07:45:30.333854 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vcpks" podStartSLOduration=2.874520672 podStartE2EDuration="5.33383041s" podCreationTimestamp="2025-10-14 07:45:25 +0000 UTC" firstStartedPulling="2025-10-14 07:45:27.283769993 +0000 UTC m=+3343.867816650" lastFinishedPulling="2025-10-14 07:45:29.743079771 +0000 UTC m=+3346.327126388" observedRunningTime="2025-10-14 07:45:30.329566899 +0000 UTC m=+3346.913613536" watchObservedRunningTime="2025-10-14 07:45:30.33383041 +0000 UTC m=+3346.917877047" Oct 14 07:45:35 crc kubenswrapper[5018]: I1014 07:45:35.708222 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:35 crc kubenswrapper[5018]: I1014 07:45:35.708927 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:35 crc kubenswrapper[5018]: I1014 07:45:35.787854 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:36 crc kubenswrapper[5018]: I1014 07:45:36.451546 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:36 crc kubenswrapper[5018]: I1014 07:45:36.931406 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:38 crc kubenswrapper[5018]: I1014 07:45:38.374378 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vcpks" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="registry-server" containerID="cri-o://cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f" gracePeriod=2 Oct 14 07:45:38 crc kubenswrapper[5018]: I1014 07:45:38.842529 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.010156 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content\") pod \"189783ca-ccdd-40c1-9ca4-a0758517f275\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.010530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4468d\" (UniqueName: \"kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d\") pod \"189783ca-ccdd-40c1-9ca4-a0758517f275\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.010779 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities\") pod \"189783ca-ccdd-40c1-9ca4-a0758517f275\" (UID: \"189783ca-ccdd-40c1-9ca4-a0758517f275\") " Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.012110 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities" (OuterVolumeSpecName: "utilities") pod "189783ca-ccdd-40c1-9ca4-a0758517f275" (UID: "189783ca-ccdd-40c1-9ca4-a0758517f275"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.020562 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d" (OuterVolumeSpecName: "kube-api-access-4468d") pod "189783ca-ccdd-40c1-9ca4-a0758517f275" (UID: "189783ca-ccdd-40c1-9ca4-a0758517f275"). InnerVolumeSpecName "kube-api-access-4468d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.113229 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4468d\" (UniqueName: \"kubernetes.io/projected/189783ca-ccdd-40c1-9ca4-a0758517f275-kube-api-access-4468d\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.113287 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.386387 5018 generic.go:334] "Generic (PLEG): container finished" podID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerID="cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f" exitCode=0 Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.386469 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerDied","Data":"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f"} Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.388194 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vcpks" event={"ID":"189783ca-ccdd-40c1-9ca4-a0758517f275","Type":"ContainerDied","Data":"fde75702ae805e72da6f0c0b1bad965017ab75d02195dd67ff8b818ffc16131f"} Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.386612 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vcpks" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.388234 5018 scope.go:117] "RemoveContainer" containerID="cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.425763 5018 scope.go:117] "RemoveContainer" containerID="fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.450212 5018 scope.go:117] "RemoveContainer" containerID="3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.494181 5018 scope.go:117] "RemoveContainer" containerID="cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f" Oct 14 07:45:39 crc kubenswrapper[5018]: E1014 07:45:39.494844 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f\": container with ID starting with cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f not found: ID does not exist" containerID="cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.494899 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f"} err="failed to get container status \"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f\": rpc error: code = NotFound desc = could not find container \"cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f\": container with ID starting with cc7978dd60cb3b693960dab840587e70ec857c9b69fe3e7d8c6850478dbc4a7f not found: ID does not exist" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.494935 5018 scope.go:117] "RemoveContainer" containerID="fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c" Oct 14 07:45:39 crc kubenswrapper[5018]: E1014 07:45:39.495358 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c\": container with ID starting with fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c not found: ID does not exist" containerID="fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.495386 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c"} err="failed to get container status \"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c\": rpc error: code = NotFound desc = could not find container \"fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c\": container with ID starting with fffb5119918b05fa0c742b6a69422d236ff1d9ab18a80b72d2fffe8dfbc5664c not found: ID does not exist" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.495402 5018 scope.go:117] "RemoveContainer" containerID="3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3" Oct 14 07:45:39 crc kubenswrapper[5018]: E1014 07:45:39.495778 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3\": container with ID starting with 3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3 not found: ID does not exist" containerID="3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.495809 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3"} err="failed to get container status \"3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3\": rpc error: code = NotFound desc = could not find container \"3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3\": container with ID starting with 3923a1da101d5e671a3f6a1ad4354e38ff3fa164cf17f0b3861f0ad5badf08a3 not found: ID does not exist" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.779379 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "189783ca-ccdd-40c1-9ca4-a0758517f275" (UID: "189783ca-ccdd-40c1-9ca4-a0758517f275"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:39 crc kubenswrapper[5018]: I1014 07:45:39.823546 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/189783ca-ccdd-40c1-9ca4-a0758517f275-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:40 crc kubenswrapper[5018]: I1014 07:45:40.026569 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:40 crc kubenswrapper[5018]: I1014 07:45:40.033298 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vcpks"] Oct 14 07:45:40 crc kubenswrapper[5018]: I1014 07:45:40.612961 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" path="/var/lib/kubelet/pods/189783ca-ccdd-40c1-9ca4-a0758517f275/volumes" Oct 14 07:46:02 crc kubenswrapper[5018]: I1014 07:46:02.463211 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:46:02 crc kubenswrapper[5018]: I1014 07:46:02.463892 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.590757 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:07 crc kubenswrapper[5018]: E1014 07:46:07.591780 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="extract-content" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.591800 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="extract-content" Oct 14 07:46:07 crc kubenswrapper[5018]: E1014 07:46:07.591818 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="extract-utilities" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.591828 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="extract-utilities" Oct 14 07:46:07 crc kubenswrapper[5018]: E1014 07:46:07.591856 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="registry-server" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.591865 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="registry-server" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.592040 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="189783ca-ccdd-40c1-9ca4-a0758517f275" containerName="registry-server" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.593580 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.603814 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.760261 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5tmn\" (UniqueName: \"kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.760357 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.760441 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.861566 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5tmn\" (UniqueName: \"kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.861753 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.861829 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.862343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.862401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.891614 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5tmn\" (UniqueName: \"kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn\") pod \"community-operators-7gp4g\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:07 crc kubenswrapper[5018]: I1014 07:46:07.939506 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:08 crc kubenswrapper[5018]: I1014 07:46:08.483851 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:08 crc kubenswrapper[5018]: I1014 07:46:08.668837 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerStarted","Data":"667687c5f390efe4ca43ff5ef436847335cbdb564eac77651c4c04b45e625917"} Oct 14 07:46:09 crc kubenswrapper[5018]: I1014 07:46:09.681133 5018 generic.go:334] "Generic (PLEG): container finished" podID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerID="8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02" exitCode=0 Oct 14 07:46:09 crc kubenswrapper[5018]: I1014 07:46:09.681228 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerDied","Data":"8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02"} Oct 14 07:46:09 crc kubenswrapper[5018]: I1014 07:46:09.686782 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:46:10 crc kubenswrapper[5018]: I1014 07:46:10.694764 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerStarted","Data":"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f"} Oct 14 07:46:11 crc kubenswrapper[5018]: I1014 07:46:11.706687 5018 generic.go:334] "Generic (PLEG): container finished" podID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerID="7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f" exitCode=0 Oct 14 07:46:11 crc kubenswrapper[5018]: I1014 07:46:11.706791 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerDied","Data":"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f"} Oct 14 07:46:12 crc kubenswrapper[5018]: I1014 07:46:12.718697 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerStarted","Data":"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1"} Oct 14 07:46:12 crc kubenswrapper[5018]: I1014 07:46:12.741061 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7gp4g" podStartSLOduration=3.270599577 podStartE2EDuration="5.741037772s" podCreationTimestamp="2025-10-14 07:46:07 +0000 UTC" firstStartedPulling="2025-10-14 07:46:09.684051341 +0000 UTC m=+3386.268098008" lastFinishedPulling="2025-10-14 07:46:12.154489566 +0000 UTC m=+3388.738536203" observedRunningTime="2025-10-14 07:46:12.737663562 +0000 UTC m=+3389.321710189" watchObservedRunningTime="2025-10-14 07:46:12.741037772 +0000 UTC m=+3389.325084409" Oct 14 07:46:17 crc kubenswrapper[5018]: I1014 07:46:17.940667 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:17 crc kubenswrapper[5018]: I1014 07:46:17.941313 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:18 crc kubenswrapper[5018]: I1014 07:46:18.021283 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:18 crc kubenswrapper[5018]: I1014 07:46:18.858212 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:18 crc kubenswrapper[5018]: I1014 07:46:18.912253 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:20 crc kubenswrapper[5018]: I1014 07:46:20.815011 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7gp4g" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="registry-server" containerID="cri-o://84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1" gracePeriod=2 Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.273217 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.388678 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities\") pod \"29c895d7-dabd-446d-aa77-e33152fe9b3e\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.388860 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5tmn\" (UniqueName: \"kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn\") pod \"29c895d7-dabd-446d-aa77-e33152fe9b3e\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.388987 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content\") pod \"29c895d7-dabd-446d-aa77-e33152fe9b3e\" (UID: \"29c895d7-dabd-446d-aa77-e33152fe9b3e\") " Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.389924 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities" (OuterVolumeSpecName: "utilities") pod "29c895d7-dabd-446d-aa77-e33152fe9b3e" (UID: "29c895d7-dabd-446d-aa77-e33152fe9b3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.395200 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn" (OuterVolumeSpecName: "kube-api-access-h5tmn") pod "29c895d7-dabd-446d-aa77-e33152fe9b3e" (UID: "29c895d7-dabd-446d-aa77-e33152fe9b3e"). InnerVolumeSpecName "kube-api-access-h5tmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.454205 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29c895d7-dabd-446d-aa77-e33152fe9b3e" (UID: "29c895d7-dabd-446d-aa77-e33152fe9b3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.490886 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5tmn\" (UniqueName: \"kubernetes.io/projected/29c895d7-dabd-446d-aa77-e33152fe9b3e-kube-api-access-h5tmn\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.490935 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.490954 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29c895d7-dabd-446d-aa77-e33152fe9b3e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.825140 5018 generic.go:334] "Generic (PLEG): container finished" podID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerID="84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1" exitCode=0 Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.825198 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerDied","Data":"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1"} Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.825228 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gp4g" event={"ID":"29c895d7-dabd-446d-aa77-e33152fe9b3e","Type":"ContainerDied","Data":"667687c5f390efe4ca43ff5ef436847335cbdb564eac77651c4c04b45e625917"} Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.825251 5018 scope.go:117] "RemoveContainer" containerID="84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.825399 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gp4g" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.848841 5018 scope.go:117] "RemoveContainer" containerID="7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.869118 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.875645 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7gp4g"] Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.909335 5018 scope.go:117] "RemoveContainer" containerID="8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.934248 5018 scope.go:117] "RemoveContainer" containerID="84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1" Oct 14 07:46:21 crc kubenswrapper[5018]: E1014 07:46:21.934719 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1\": container with ID starting with 84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1 not found: ID does not exist" containerID="84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.934774 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1"} err="failed to get container status \"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1\": rpc error: code = NotFound desc = could not find container \"84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1\": container with ID starting with 84ca80c7eb40eac9b26a957c35df0fac08e76053bae4e4bf5a6582d55d3a1bf1 not found: ID does not exist" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.934811 5018 scope.go:117] "RemoveContainer" containerID="7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f" Oct 14 07:46:21 crc kubenswrapper[5018]: E1014 07:46:21.935134 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f\": container with ID starting with 7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f not found: ID does not exist" containerID="7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.935233 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f"} err="failed to get container status \"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f\": rpc error: code = NotFound desc = could not find container \"7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f\": container with ID starting with 7c78dca71640861bfa4d2afa2a87d4e577488be098ad4e32b36dd65115c5dd1f not found: ID does not exist" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.935263 5018 scope.go:117] "RemoveContainer" containerID="8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02" Oct 14 07:46:21 crc kubenswrapper[5018]: E1014 07:46:21.935848 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02\": container with ID starting with 8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02 not found: ID does not exist" containerID="8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02" Oct 14 07:46:21 crc kubenswrapper[5018]: I1014 07:46:21.935930 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02"} err="failed to get container status \"8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02\": rpc error: code = NotFound desc = could not find container \"8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02\": container with ID starting with 8cf4a609e8fd8926e1e592dd8b444ccda834f390e428c433c8714df950f36b02 not found: ID does not exist" Oct 14 07:46:22 crc kubenswrapper[5018]: I1014 07:46:22.621816 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" path="/var/lib/kubelet/pods/29c895d7-dabd-446d-aa77-e33152fe9b3e/volumes" Oct 14 07:46:32 crc kubenswrapper[5018]: I1014 07:46:32.464174 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:46:32 crc kubenswrapper[5018]: I1014 07:46:32.465102 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:47:02 crc kubenswrapper[5018]: I1014 07:47:02.463904 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:47:02 crc kubenswrapper[5018]: I1014 07:47:02.464725 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:47:02 crc kubenswrapper[5018]: I1014 07:47:02.464794 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:47:02 crc kubenswrapper[5018]: I1014 07:47:02.465685 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:47:02 crc kubenswrapper[5018]: I1014 07:47:02.465810 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a" gracePeriod=600 Oct 14 07:47:03 crc kubenswrapper[5018]: I1014 07:47:03.246152 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a" exitCode=0 Oct 14 07:47:03 crc kubenswrapper[5018]: I1014 07:47:03.246254 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a"} Oct 14 07:47:03 crc kubenswrapper[5018]: I1014 07:47:03.246641 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed"} Oct 14 07:47:03 crc kubenswrapper[5018]: I1014 07:47:03.246672 5018 scope.go:117] "RemoveContainer" containerID="6da7c4182627e384548051b10596a5eb66e5bd673ebf284972bb58b6c3669212" Oct 14 07:49:02 crc kubenswrapper[5018]: I1014 07:49:02.463731 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:49:02 crc kubenswrapper[5018]: I1014 07:49:02.464370 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:49:32 crc kubenswrapper[5018]: I1014 07:49:32.463868 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:49:32 crc kubenswrapper[5018]: I1014 07:49:32.464681 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:50:02 crc kubenswrapper[5018]: I1014 07:50:02.464009 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:50:02 crc kubenswrapper[5018]: I1014 07:50:02.464840 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:50:02 crc kubenswrapper[5018]: I1014 07:50:02.464919 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:50:02 crc kubenswrapper[5018]: I1014 07:50:02.465907 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:50:02 crc kubenswrapper[5018]: I1014 07:50:02.466007 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" gracePeriod=600 Oct 14 07:50:02 crc kubenswrapper[5018]: E1014 07:50:02.616940 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:50:03 crc kubenswrapper[5018]: I1014 07:50:03.096362 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" exitCode=0 Oct 14 07:50:03 crc kubenswrapper[5018]: I1014 07:50:03.096433 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed"} Oct 14 07:50:03 crc kubenswrapper[5018]: I1014 07:50:03.096484 5018 scope.go:117] "RemoveContainer" containerID="2f735fe2d3afa86b996206d4fae2a7346b21283f4c5ed2a7a964541ca6dcef3a" Oct 14 07:50:03 crc kubenswrapper[5018]: I1014 07:50:03.097144 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:50:03 crc kubenswrapper[5018]: E1014 07:50:03.097590 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:50:14 crc kubenswrapper[5018]: I1014 07:50:14.618277 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:50:14 crc kubenswrapper[5018]: E1014 07:50:14.622667 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:50:29 crc kubenswrapper[5018]: I1014 07:50:29.605550 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:50:29 crc kubenswrapper[5018]: E1014 07:50:29.606733 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:50:42 crc kubenswrapper[5018]: I1014 07:50:42.605510 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:50:42 crc kubenswrapper[5018]: E1014 07:50:42.606329 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:50:53 crc kubenswrapper[5018]: I1014 07:50:53.605831 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:50:53 crc kubenswrapper[5018]: E1014 07:50:53.606905 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:51:04 crc kubenswrapper[5018]: I1014 07:51:04.613218 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:51:04 crc kubenswrapper[5018]: E1014 07:51:04.614373 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:51:17 crc kubenswrapper[5018]: I1014 07:51:17.605311 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:51:17 crc kubenswrapper[5018]: E1014 07:51:17.606474 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:51:29 crc kubenswrapper[5018]: I1014 07:51:29.604964 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:51:29 crc kubenswrapper[5018]: E1014 07:51:29.605992 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:51:43 crc kubenswrapper[5018]: I1014 07:51:43.605266 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:51:43 crc kubenswrapper[5018]: E1014 07:51:43.606084 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:51:57 crc kubenswrapper[5018]: I1014 07:51:57.605794 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:51:57 crc kubenswrapper[5018]: E1014 07:51:57.606759 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:52:11 crc kubenswrapper[5018]: I1014 07:52:11.605207 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:52:11 crc kubenswrapper[5018]: E1014 07:52:11.607308 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:52:23 crc kubenswrapper[5018]: I1014 07:52:23.605328 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:52:23 crc kubenswrapper[5018]: E1014 07:52:23.606277 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:52:37 crc kubenswrapper[5018]: I1014 07:52:37.606016 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:52:37 crc kubenswrapper[5018]: E1014 07:52:37.607451 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:52:48 crc kubenswrapper[5018]: I1014 07:52:48.604716 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:52:48 crc kubenswrapper[5018]: E1014 07:52:48.605671 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.487301 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:02 crc kubenswrapper[5018]: E1014 07:53:02.488266 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="extract-content" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.488283 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="extract-content" Oct 14 07:53:02 crc kubenswrapper[5018]: E1014 07:53:02.488299 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="registry-server" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.488306 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="registry-server" Oct 14 07:53:02 crc kubenswrapper[5018]: E1014 07:53:02.488314 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="extract-utilities" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.488322 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="extract-utilities" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.488517 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c895d7-dabd-446d-aa77-e33152fe9b3e" containerName="registry-server" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.489654 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.515216 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.605326 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:53:02 crc kubenswrapper[5018]: E1014 07:53:02.606226 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.676691 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.676753 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.676869 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdjxr\" (UniqueName: \"kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.777784 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdjxr\" (UniqueName: \"kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.777849 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.777875 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.778295 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.778565 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.800863 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdjxr\" (UniqueName: \"kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr\") pod \"redhat-marketplace-h5mt6\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:02 crc kubenswrapper[5018]: I1014 07:53:02.857139 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:03 crc kubenswrapper[5018]: I1014 07:53:03.310887 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:03 crc kubenswrapper[5018]: I1014 07:53:03.782703 5018 generic.go:334] "Generic (PLEG): container finished" podID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerID="de53f58a4769ae9b65f9bd94e55e912f5d527624d27f2f62bacd1bff3990560f" exitCode=0 Oct 14 07:53:03 crc kubenswrapper[5018]: I1014 07:53:03.782825 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerDied","Data":"de53f58a4769ae9b65f9bd94e55e912f5d527624d27f2f62bacd1bff3990560f"} Oct 14 07:53:03 crc kubenswrapper[5018]: I1014 07:53:03.782871 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerStarted","Data":"b81ac217f1d963a6af1c6f2b68097f1e459132b2e397c23911a3abf9684725d7"} Oct 14 07:53:03 crc kubenswrapper[5018]: I1014 07:53:03.786967 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:53:05 crc kubenswrapper[5018]: I1014 07:53:05.804296 5018 generic.go:334] "Generic (PLEG): container finished" podID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerID="74e2c3deb2cba626b48fbfc8d067b4002d9d324c11bace6e3e88e55898d5cbe9" exitCode=0 Oct 14 07:53:05 crc kubenswrapper[5018]: I1014 07:53:05.804394 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerDied","Data":"74e2c3deb2cba626b48fbfc8d067b4002d9d324c11bace6e3e88e55898d5cbe9"} Oct 14 07:53:06 crc kubenswrapper[5018]: I1014 07:53:06.816917 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerStarted","Data":"8013fe882cdc897e8b3fc977e39e1f617f308b85fde378ff88b9f548a2b146fd"} Oct 14 07:53:06 crc kubenswrapper[5018]: I1014 07:53:06.837241 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5mt6" podStartSLOduration=2.132459073 podStartE2EDuration="4.837217703s" podCreationTimestamp="2025-10-14 07:53:02 +0000 UTC" firstStartedPulling="2025-10-14 07:53:03.78658136 +0000 UTC m=+3800.370628017" lastFinishedPulling="2025-10-14 07:53:06.49134002 +0000 UTC m=+3803.075386647" observedRunningTime="2025-10-14 07:53:06.832435718 +0000 UTC m=+3803.416482345" watchObservedRunningTime="2025-10-14 07:53:06.837217703 +0000 UTC m=+3803.421264350" Oct 14 07:53:12 crc kubenswrapper[5018]: I1014 07:53:12.858226 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:12 crc kubenswrapper[5018]: I1014 07:53:12.858792 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:12 crc kubenswrapper[5018]: I1014 07:53:12.919551 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:12 crc kubenswrapper[5018]: I1014 07:53:12.959684 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:13 crc kubenswrapper[5018]: I1014 07:53:13.155944 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:14 crc kubenswrapper[5018]: I1014 07:53:14.878238 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h5mt6" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="registry-server" containerID="cri-o://8013fe882cdc897e8b3fc977e39e1f617f308b85fde378ff88b9f548a2b146fd" gracePeriod=2 Oct 14 07:53:15 crc kubenswrapper[5018]: I1014 07:53:15.888125 5018 generic.go:334] "Generic (PLEG): container finished" podID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerID="8013fe882cdc897e8b3fc977e39e1f617f308b85fde378ff88b9f548a2b146fd" exitCode=0 Oct 14 07:53:15 crc kubenswrapper[5018]: I1014 07:53:15.888205 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerDied","Data":"8013fe882cdc897e8b3fc977e39e1f617f308b85fde378ff88b9f548a2b146fd"} Oct 14 07:53:15 crc kubenswrapper[5018]: I1014 07:53:15.888499 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5mt6" event={"ID":"05f7c115-5fd1-48f1-82bb-57e4f9ac9599","Type":"ContainerDied","Data":"b81ac217f1d963a6af1c6f2b68097f1e459132b2e397c23911a3abf9684725d7"} Oct 14 07:53:15 crc kubenswrapper[5018]: I1014 07:53:15.888516 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b81ac217f1d963a6af1c6f2b68097f1e459132b2e397c23911a3abf9684725d7" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.109950 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.286116 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdjxr\" (UniqueName: \"kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr\") pod \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.286167 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities\") pod \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.286213 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content\") pod \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\" (UID: \"05f7c115-5fd1-48f1-82bb-57e4f9ac9599\") " Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.287159 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities" (OuterVolumeSpecName: "utilities") pod "05f7c115-5fd1-48f1-82bb-57e4f9ac9599" (UID: "05f7c115-5fd1-48f1-82bb-57e4f9ac9599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.294714 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr" (OuterVolumeSpecName: "kube-api-access-gdjxr") pod "05f7c115-5fd1-48f1-82bb-57e4f9ac9599" (UID: "05f7c115-5fd1-48f1-82bb-57e4f9ac9599"). InnerVolumeSpecName "kube-api-access-gdjxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.299454 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05f7c115-5fd1-48f1-82bb-57e4f9ac9599" (UID: "05f7c115-5fd1-48f1-82bb-57e4f9ac9599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.388667 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdjxr\" (UniqueName: \"kubernetes.io/projected/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-kube-api-access-gdjxr\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.389004 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.389024 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f7c115-5fd1-48f1-82bb-57e4f9ac9599-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.604551 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:53:16 crc kubenswrapper[5018]: E1014 07:53:16.604958 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.895541 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5mt6" Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.919026 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:16 crc kubenswrapper[5018]: I1014 07:53:16.924759 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5mt6"] Oct 14 07:53:18 crc kubenswrapper[5018]: I1014 07:53:18.619941 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" path="/var/lib/kubelet/pods/05f7c115-5fd1-48f1-82bb-57e4f9ac9599/volumes" Oct 14 07:53:31 crc kubenswrapper[5018]: I1014 07:53:31.604954 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:53:31 crc kubenswrapper[5018]: E1014 07:53:31.605965 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:53:44 crc kubenswrapper[5018]: I1014 07:53:44.618568 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:53:44 crc kubenswrapper[5018]: E1014 07:53:44.619727 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:53:57 crc kubenswrapper[5018]: I1014 07:53:57.605807 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:53:57 crc kubenswrapper[5018]: E1014 07:53:57.609833 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:54:12 crc kubenswrapper[5018]: I1014 07:54:12.605976 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:54:12 crc kubenswrapper[5018]: E1014 07:54:12.607356 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:54:26 crc kubenswrapper[5018]: I1014 07:54:26.604994 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:54:26 crc kubenswrapper[5018]: E1014 07:54:26.605672 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:54:37 crc kubenswrapper[5018]: I1014 07:54:37.604978 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:54:37 crc kubenswrapper[5018]: E1014 07:54:37.605925 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:54:52 crc kubenswrapper[5018]: I1014 07:54:52.605053 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:54:52 crc kubenswrapper[5018]: E1014 07:54:52.606456 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 07:55:03 crc kubenswrapper[5018]: I1014 07:55:03.605535 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:55:03 crc kubenswrapper[5018]: I1014 07:55:03.886407 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830"} Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.579032 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:40 crc kubenswrapper[5018]: E1014 07:55:40.579941 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="registry-server" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.579957 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="registry-server" Oct 14 07:55:40 crc kubenswrapper[5018]: E1014 07:55:40.579974 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="extract-utilities" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.579982 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="extract-utilities" Oct 14 07:55:40 crc kubenswrapper[5018]: E1014 07:55:40.580013 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="extract-content" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.580020 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="extract-content" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.580201 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f7c115-5fd1-48f1-82bb-57e4f9ac9599" containerName="registry-server" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.581333 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.649222 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.676121 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.676206 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtvnw\" (UniqueName: \"kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.676696 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.778333 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.778439 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtvnw\" (UniqueName: \"kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.778556 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.779094 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.779249 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.801103 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtvnw\" (UniqueName: \"kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw\") pod \"certified-operators-5l8hv\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:40 crc kubenswrapper[5018]: I1014 07:55:40.903780 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:41 crc kubenswrapper[5018]: I1014 07:55:41.166241 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:41 crc kubenswrapper[5018]: I1014 07:55:41.229229 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerStarted","Data":"37a941089972796fc015c41891b22bdfd665dabca59a3897371eb971fc67fbff"} Oct 14 07:55:42 crc kubenswrapper[5018]: I1014 07:55:42.242785 5018 generic.go:334] "Generic (PLEG): container finished" podID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerID="60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0" exitCode=0 Oct 14 07:55:42 crc kubenswrapper[5018]: I1014 07:55:42.242837 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerDied","Data":"60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0"} Oct 14 07:55:43 crc kubenswrapper[5018]: I1014 07:55:43.252019 5018 generic.go:334] "Generic (PLEG): container finished" podID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerID="dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45" exitCode=0 Oct 14 07:55:43 crc kubenswrapper[5018]: I1014 07:55:43.252259 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerDied","Data":"dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45"} Oct 14 07:55:44 crc kubenswrapper[5018]: I1014 07:55:44.262149 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerStarted","Data":"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb"} Oct 14 07:55:44 crc kubenswrapper[5018]: I1014 07:55:44.291994 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5l8hv" podStartSLOduration=2.819555278 podStartE2EDuration="4.291966305s" podCreationTimestamp="2025-10-14 07:55:40 +0000 UTC" firstStartedPulling="2025-10-14 07:55:42.245139206 +0000 UTC m=+3958.829185833" lastFinishedPulling="2025-10-14 07:55:43.717550233 +0000 UTC m=+3960.301596860" observedRunningTime="2025-10-14 07:55:44.287120386 +0000 UTC m=+3960.871167043" watchObservedRunningTime="2025-10-14 07:55:44.291966305 +0000 UTC m=+3960.876012952" Oct 14 07:55:50 crc kubenswrapper[5018]: I1014 07:55:50.904378 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:50 crc kubenswrapper[5018]: I1014 07:55:50.905069 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:50 crc kubenswrapper[5018]: I1014 07:55:50.956173 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:51 crc kubenswrapper[5018]: I1014 07:55:51.617794 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:51 crc kubenswrapper[5018]: I1014 07:55:51.672705 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.336702 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5l8hv" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="registry-server" containerID="cri-o://b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb" gracePeriod=2 Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.857947 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.982086 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtvnw\" (UniqueName: \"kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw\") pod \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.982183 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities\") pod \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.982210 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content\") pod \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\" (UID: \"b13a9b7d-f738-4d0e-982c-a6ff2b1df413\") " Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.983420 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities" (OuterVolumeSpecName: "utilities") pod "b13a9b7d-f738-4d0e-982c-a6ff2b1df413" (UID: "b13a9b7d-f738-4d0e-982c-a6ff2b1df413"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:53 crc kubenswrapper[5018]: I1014 07:55:53.990892 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw" (OuterVolumeSpecName: "kube-api-access-rtvnw") pod "b13a9b7d-f738-4d0e-982c-a6ff2b1df413" (UID: "b13a9b7d-f738-4d0e-982c-a6ff2b1df413"). InnerVolumeSpecName "kube-api-access-rtvnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.042498 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b13a9b7d-f738-4d0e-982c-a6ff2b1df413" (UID: "b13a9b7d-f738-4d0e-982c-a6ff2b1df413"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.083326 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtvnw\" (UniqueName: \"kubernetes.io/projected/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-kube-api-access-rtvnw\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.083360 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.083372 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b13a9b7d-f738-4d0e-982c-a6ff2b1df413-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.361036 5018 generic.go:334] "Generic (PLEG): container finished" podID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerID="b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb" exitCode=0 Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.361087 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerDied","Data":"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb"} Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.361127 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5l8hv" event={"ID":"b13a9b7d-f738-4d0e-982c-a6ff2b1df413","Type":"ContainerDied","Data":"37a941089972796fc015c41891b22bdfd665dabca59a3897371eb971fc67fbff"} Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.361149 5018 scope.go:117] "RemoveContainer" containerID="b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.361178 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5l8hv" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.390997 5018 scope.go:117] "RemoveContainer" containerID="dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.420776 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.425057 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5l8hv"] Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.432368 5018 scope.go:117] "RemoveContainer" containerID="60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.461110 5018 scope.go:117] "RemoveContainer" containerID="b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb" Oct 14 07:55:54 crc kubenswrapper[5018]: E1014 07:55:54.461589 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb\": container with ID starting with b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb not found: ID does not exist" containerID="b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.461649 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb"} err="failed to get container status \"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb\": rpc error: code = NotFound desc = could not find container \"b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb\": container with ID starting with b737350f4b6382f4a03f6b4a75864f807b419a16d473925f60cf215bf2cebdeb not found: ID does not exist" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.461673 5018 scope.go:117] "RemoveContainer" containerID="dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45" Oct 14 07:55:54 crc kubenswrapper[5018]: E1014 07:55:54.462113 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45\": container with ID starting with dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45 not found: ID does not exist" containerID="dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.462170 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45"} err="failed to get container status \"dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45\": rpc error: code = NotFound desc = could not find container \"dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45\": container with ID starting with dbcc04b8a489d1643f521a0e3dc9bb11aff66cd8ef89e4ff1d074e79680cea45 not found: ID does not exist" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.462202 5018 scope.go:117] "RemoveContainer" containerID="60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0" Oct 14 07:55:54 crc kubenswrapper[5018]: E1014 07:55:54.462550 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0\": container with ID starting with 60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0 not found: ID does not exist" containerID="60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.462588 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0"} err="failed to get container status \"60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0\": rpc error: code = NotFound desc = could not find container \"60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0\": container with ID starting with 60bc5146e2e3a2226ca26c28b3ac0ad600bcec15db46505a96e5a97151b9eff0 not found: ID does not exist" Oct 14 07:55:54 crc kubenswrapper[5018]: I1014 07:55:54.615545 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" path="/var/lib/kubelet/pods/b13a9b7d-f738-4d0e-982c-a6ff2b1df413/volumes" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.886765 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:12 crc kubenswrapper[5018]: E1014 07:56:12.887642 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="extract-utilities" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.887657 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="extract-utilities" Oct 14 07:56:12 crc kubenswrapper[5018]: E1014 07:56:12.887672 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="registry-server" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.887679 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="registry-server" Oct 14 07:56:12 crc kubenswrapper[5018]: E1014 07:56:12.887718 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="extract-content" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.887727 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="extract-content" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.887882 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13a9b7d-f738-4d0e-982c-a6ff2b1df413" containerName="registry-server" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.889075 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.908510 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.967101 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.967434 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:12 crc kubenswrapper[5018]: I1014 07:56:12.967572 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzwss\" (UniqueName: \"kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.068830 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.068896 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.068938 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzwss\" (UniqueName: \"kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.069437 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.069533 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.093422 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzwss\" (UniqueName: \"kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss\") pod \"redhat-operators-kw2jk\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.212229 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:13 crc kubenswrapper[5018]: I1014 07:56:13.656100 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:14 crc kubenswrapper[5018]: I1014 07:56:14.546750 5018 generic.go:334] "Generic (PLEG): container finished" podID="92829142-b5b3-4262-8141-8ad0de90756c" containerID="58aa0b183fe2cc865b915a467dac4a69d3f18e0aec3c5d688df67221f5bbf929" exitCode=0 Oct 14 07:56:14 crc kubenswrapper[5018]: I1014 07:56:14.546827 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerDied","Data":"58aa0b183fe2cc865b915a467dac4a69d3f18e0aec3c5d688df67221f5bbf929"} Oct 14 07:56:14 crc kubenswrapper[5018]: I1014 07:56:14.547126 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerStarted","Data":"5a0b8c6b9d906a6c87816976228a677b0493efcf0e02144a4612caf23b59b47f"} Oct 14 07:56:15 crc kubenswrapper[5018]: I1014 07:56:15.556477 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerStarted","Data":"4ac01113441e2db04d42a42e20d0b8235719860d96dffa8bf688261f65044674"} Oct 14 07:56:16 crc kubenswrapper[5018]: I1014 07:56:16.569011 5018 generic.go:334] "Generic (PLEG): container finished" podID="92829142-b5b3-4262-8141-8ad0de90756c" containerID="4ac01113441e2db04d42a42e20d0b8235719860d96dffa8bf688261f65044674" exitCode=0 Oct 14 07:56:16 crc kubenswrapper[5018]: I1014 07:56:16.569124 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerDied","Data":"4ac01113441e2db04d42a42e20d0b8235719860d96dffa8bf688261f65044674"} Oct 14 07:56:17 crc kubenswrapper[5018]: I1014 07:56:17.579330 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerStarted","Data":"706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b"} Oct 14 07:56:17 crc kubenswrapper[5018]: I1014 07:56:17.607023 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kw2jk" podStartSLOduration=3.125295802 podStartE2EDuration="5.606994903s" podCreationTimestamp="2025-10-14 07:56:12 +0000 UTC" firstStartedPulling="2025-10-14 07:56:14.548999317 +0000 UTC m=+3991.133045984" lastFinishedPulling="2025-10-14 07:56:17.030698428 +0000 UTC m=+3993.614745085" observedRunningTime="2025-10-14 07:56:17.599583721 +0000 UTC m=+3994.183630398" watchObservedRunningTime="2025-10-14 07:56:17.606994903 +0000 UTC m=+3994.191041560" Oct 14 07:56:23 crc kubenswrapper[5018]: I1014 07:56:23.212430 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:23 crc kubenswrapper[5018]: I1014 07:56:23.212833 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:23 crc kubenswrapper[5018]: I1014 07:56:23.282969 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:23 crc kubenswrapper[5018]: I1014 07:56:23.666856 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:23 crc kubenswrapper[5018]: I1014 07:56:23.714223 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:25 crc kubenswrapper[5018]: I1014 07:56:25.655841 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kw2jk" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="registry-server" containerID="cri-o://706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b" gracePeriod=2 Oct 14 07:56:26 crc kubenswrapper[5018]: E1014 07:56:26.165292 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92829142_b5b3_4262_8141_8ad0de90756c.slice/crio-conmon-706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.665480 5018 generic.go:334] "Generic (PLEG): container finished" podID="92829142-b5b3-4262-8141-8ad0de90756c" containerID="706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b" exitCode=0 Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.665540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerDied","Data":"706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b"} Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.866846 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.978352 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzwss\" (UniqueName: \"kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss\") pod \"92829142-b5b3-4262-8141-8ad0de90756c\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.978532 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities\") pod \"92829142-b5b3-4262-8141-8ad0de90756c\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.978708 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content\") pod \"92829142-b5b3-4262-8141-8ad0de90756c\" (UID: \"92829142-b5b3-4262-8141-8ad0de90756c\") " Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.980098 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities" (OuterVolumeSpecName: "utilities") pod "92829142-b5b3-4262-8141-8ad0de90756c" (UID: "92829142-b5b3-4262-8141-8ad0de90756c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:26 crc kubenswrapper[5018]: I1014 07:56:26.984576 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss" (OuterVolumeSpecName: "kube-api-access-tzwss") pod "92829142-b5b3-4262-8141-8ad0de90756c" (UID: "92829142-b5b3-4262-8141-8ad0de90756c"). InnerVolumeSpecName "kube-api-access-tzwss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.070390 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92829142-b5b3-4262-8141-8ad0de90756c" (UID: "92829142-b5b3-4262-8141-8ad0de90756c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.081213 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzwss\" (UniqueName: \"kubernetes.io/projected/92829142-b5b3-4262-8141-8ad0de90756c-kube-api-access-tzwss\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.081265 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.081381 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92829142-b5b3-4262-8141-8ad0de90756c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.679112 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw2jk" event={"ID":"92829142-b5b3-4262-8141-8ad0de90756c","Type":"ContainerDied","Data":"5a0b8c6b9d906a6c87816976228a677b0493efcf0e02144a4612caf23b59b47f"} Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.679200 5018 scope.go:117] "RemoveContainer" containerID="706d874ac09b412e25223d53657138c6179f78680a99ebec11e8e84a32bbe96b" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.679390 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw2jk" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.711383 5018 scope.go:117] "RemoveContainer" containerID="4ac01113441e2db04d42a42e20d0b8235719860d96dffa8bf688261f65044674" Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.733771 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.742379 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kw2jk"] Oct 14 07:56:27 crc kubenswrapper[5018]: I1014 07:56:27.758826 5018 scope.go:117] "RemoveContainer" containerID="58aa0b183fe2cc865b915a467dac4a69d3f18e0aec3c5d688df67221f5bbf929" Oct 14 07:56:28 crc kubenswrapper[5018]: I1014 07:56:28.617671 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92829142-b5b3-4262-8141-8ad0de90756c" path="/var/lib/kubelet/pods/92829142-b5b3-4262-8141-8ad0de90756c/volumes" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.156634 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:56:56 crc kubenswrapper[5018]: E1014 07:56:56.157535 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="extract-utilities" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.157547 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="extract-utilities" Oct 14 07:56:56 crc kubenswrapper[5018]: E1014 07:56:56.157566 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="registry-server" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.157573 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="registry-server" Oct 14 07:56:56 crc kubenswrapper[5018]: E1014 07:56:56.157585 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="extract-content" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.157591 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="extract-content" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.157750 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="92829142-b5b3-4262-8141-8ad0de90756c" containerName="registry-server" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.158741 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.174982 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.206051 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8slfx\" (UniqueName: \"kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.206106 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.206136 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.307245 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8slfx\" (UniqueName: \"kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.307315 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.307342 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.308002 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.308077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.339725 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8slfx\" (UniqueName: \"kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx\") pod \"community-operators-6x2xq\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.480278 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.914871 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:56:56 crc kubenswrapper[5018]: I1014 07:56:56.937112 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerStarted","Data":"9855909994f8ff9c4c5f27366b3b5dee3f028881f5ed58f92a7262ba93cc988b"} Oct 14 07:56:57 crc kubenswrapper[5018]: I1014 07:56:57.950912 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5431865-1e72-49d4-812e-632185062568" containerID="ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c" exitCode=0 Oct 14 07:56:57 crc kubenswrapper[5018]: I1014 07:56:57.950994 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerDied","Data":"ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c"} Oct 14 07:56:58 crc kubenswrapper[5018]: I1014 07:56:58.961961 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerStarted","Data":"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc"} Oct 14 07:56:59 crc kubenswrapper[5018]: I1014 07:56:59.969329 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5431865-1e72-49d4-812e-632185062568" containerID="47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc" exitCode=0 Oct 14 07:56:59 crc kubenswrapper[5018]: I1014 07:56:59.969380 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerDied","Data":"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc"} Oct 14 07:57:00 crc kubenswrapper[5018]: I1014 07:57:00.984496 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerStarted","Data":"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c"} Oct 14 07:57:01 crc kubenswrapper[5018]: I1014 07:57:01.004873 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6x2xq" podStartSLOduration=2.199526057 podStartE2EDuration="5.004842843s" podCreationTimestamp="2025-10-14 07:56:56 +0000 UTC" firstStartedPulling="2025-10-14 07:56:57.953023304 +0000 UTC m=+4034.537069931" lastFinishedPulling="2025-10-14 07:57:00.75834009 +0000 UTC m=+4037.342386717" observedRunningTime="2025-10-14 07:57:01.002606969 +0000 UTC m=+4037.586653596" watchObservedRunningTime="2025-10-14 07:57:01.004842843 +0000 UTC m=+4037.588889500" Oct 14 07:57:06 crc kubenswrapper[5018]: I1014 07:57:06.480756 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:06 crc kubenswrapper[5018]: I1014 07:57:06.482100 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:06 crc kubenswrapper[5018]: I1014 07:57:06.552856 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:07 crc kubenswrapper[5018]: I1014 07:57:07.080698 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.345966 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.346638 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6x2xq" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="registry-server" containerID="cri-o://56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c" gracePeriod=2 Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.825748 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.919328 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities\") pod \"d5431865-1e72-49d4-812e-632185062568\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.919461 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content\") pod \"d5431865-1e72-49d4-812e-632185062568\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.919490 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8slfx\" (UniqueName: \"kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx\") pod \"d5431865-1e72-49d4-812e-632185062568\" (UID: \"d5431865-1e72-49d4-812e-632185062568\") " Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.920688 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities" (OuterVolumeSpecName: "utilities") pod "d5431865-1e72-49d4-812e-632185062568" (UID: "d5431865-1e72-49d4-812e-632185062568"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.928260 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx" (OuterVolumeSpecName: "kube-api-access-8slfx") pod "d5431865-1e72-49d4-812e-632185062568" (UID: "d5431865-1e72-49d4-812e-632185062568"). InnerVolumeSpecName "kube-api-access-8slfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:57:10 crc kubenswrapper[5018]: I1014 07:57:10.972278 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5431865-1e72-49d4-812e-632185062568" (UID: "d5431865-1e72-49d4-812e-632185062568"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.021601 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8slfx\" (UniqueName: \"kubernetes.io/projected/d5431865-1e72-49d4-812e-632185062568-kube-api-access-8slfx\") on node \"crc\" DevicePath \"\"" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.021656 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.021669 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5431865-1e72-49d4-812e-632185062568-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.084021 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5431865-1e72-49d4-812e-632185062568" containerID="56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c" exitCode=0 Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.084069 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerDied","Data":"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c"} Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.084104 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6x2xq" event={"ID":"d5431865-1e72-49d4-812e-632185062568","Type":"ContainerDied","Data":"9855909994f8ff9c4c5f27366b3b5dee3f028881f5ed58f92a7262ba93cc988b"} Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.084126 5018 scope.go:117] "RemoveContainer" containerID="56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.084129 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6x2xq" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.109231 5018 scope.go:117] "RemoveContainer" containerID="47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.119838 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.125971 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6x2xq"] Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.134317 5018 scope.go:117] "RemoveContainer" containerID="ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.180900 5018 scope.go:117] "RemoveContainer" containerID="56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c" Oct 14 07:57:11 crc kubenswrapper[5018]: E1014 07:57:11.181690 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c\": container with ID starting with 56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c not found: ID does not exist" containerID="56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.181745 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c"} err="failed to get container status \"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c\": rpc error: code = NotFound desc = could not find container \"56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c\": container with ID starting with 56167c44facdeadd76f4e2d07049af9b6a5cd77848d92e53c11fa35449751a8c not found: ID does not exist" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.181778 5018 scope.go:117] "RemoveContainer" containerID="47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc" Oct 14 07:57:11 crc kubenswrapper[5018]: E1014 07:57:11.182526 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc\": container with ID starting with 47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc not found: ID does not exist" containerID="47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.182579 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc"} err="failed to get container status \"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc\": rpc error: code = NotFound desc = could not find container \"47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc\": container with ID starting with 47c7b88c4516913f37d8cb3734f45b1bd8bda17f533749d08568d738fd48e6fc not found: ID does not exist" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.182615 5018 scope.go:117] "RemoveContainer" containerID="ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c" Oct 14 07:57:11 crc kubenswrapper[5018]: E1014 07:57:11.183335 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c\": container with ID starting with ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c not found: ID does not exist" containerID="ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c" Oct 14 07:57:11 crc kubenswrapper[5018]: I1014 07:57:11.183389 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c"} err="failed to get container status \"ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c\": rpc error: code = NotFound desc = could not find container \"ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c\": container with ID starting with ac3fb16a06484d6346f4e5069ac8e99b95da17ea717f355b9398b7b4aa70022c not found: ID does not exist" Oct 14 07:57:12 crc kubenswrapper[5018]: I1014 07:57:12.623924 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5431865-1e72-49d4-812e-632185062568" path="/var/lib/kubelet/pods/d5431865-1e72-49d4-812e-632185062568/volumes" Oct 14 07:57:32 crc kubenswrapper[5018]: I1014 07:57:32.463374 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:57:32 crc kubenswrapper[5018]: I1014 07:57:32.464175 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:58:02 crc kubenswrapper[5018]: I1014 07:58:02.463366 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:58:02 crc kubenswrapper[5018]: I1014 07:58:02.464112 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.463112 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.463740 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.463802 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.464577 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.464709 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830" gracePeriod=600 Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.808350 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830" exitCode=0 Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.808408 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830"} Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.808453 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b"} Oct 14 07:58:32 crc kubenswrapper[5018]: I1014 07:58:32.808475 5018 scope.go:117] "RemoveContainer" containerID="e2469074dab5db0e42dafa6e03ffca592c61be0917bbcc83c9ef14f25138a7ed" Oct 14 07:59:04 crc kubenswrapper[5018]: I1014 07:59:04.998159 5018 scope.go:117] "RemoveContainer" containerID="de53f58a4769ae9b65f9bd94e55e912f5d527624d27f2f62bacd1bff3990560f" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.167867 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6"] Oct 14 08:00:00 crc kubenswrapper[5018]: E1014 08:00:00.169266 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="extract-utilities" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.169302 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="extract-utilities" Oct 14 08:00:00 crc kubenswrapper[5018]: E1014 08:00:00.169346 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.169363 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[5018]: E1014 08:00:00.169427 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="extract-content" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.169447 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="extract-content" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.169821 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5431865-1e72-49d4-812e-632185062568" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.170857 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.176361 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.176375 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.187322 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6"] Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.266632 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.266848 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.266950 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxznt\" (UniqueName: \"kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.368225 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.368308 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.368338 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxznt\" (UniqueName: \"kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.369336 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.378577 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.402155 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxznt\" (UniqueName: \"kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt\") pod \"collect-profiles-29340480-962q6\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.516857 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:00 crc kubenswrapper[5018]: I1014 08:00:00.812640 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6"] Oct 14 08:00:01 crc kubenswrapper[5018]: I1014 08:00:01.706039 5018 generic.go:334] "Generic (PLEG): container finished" podID="688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" containerID="045b55da320689ea87c32d82e821d8f765d4462d588ed2727b25b680b10f83c9" exitCode=0 Oct 14 08:00:01 crc kubenswrapper[5018]: I1014 08:00:01.706157 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" event={"ID":"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8","Type":"ContainerDied","Data":"045b55da320689ea87c32d82e821d8f765d4462d588ed2727b25b680b10f83c9"} Oct 14 08:00:01 crc kubenswrapper[5018]: I1014 08:00:01.706593 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" event={"ID":"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8","Type":"ContainerStarted","Data":"588b97ac4c78bc2adc23c3418b2f1b7f3b8c0146c01099266d14620c807319b3"} Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.105377 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.209935 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume\") pod \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.210342 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxznt\" (UniqueName: \"kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt\") pod \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.210608 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume\") pod \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\" (UID: \"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8\") " Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.211439 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" (UID: "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.218850 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt" (OuterVolumeSpecName: "kube-api-access-gxznt") pod "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" (UID: "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8"). InnerVolumeSpecName "kube-api-access-gxznt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.219488 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" (UID: "688bb2d0-e6e4-495f-aa22-5dc6f890c7b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.313151 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.313226 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxznt\" (UniqueName: \"kubernetes.io/projected/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-kube-api-access-gxznt\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.313260 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.734193 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" event={"ID":"688bb2d0-e6e4-495f-aa22-5dc6f890c7b8","Type":"ContainerDied","Data":"588b97ac4c78bc2adc23c3418b2f1b7f3b8c0146c01099266d14620c807319b3"} Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.734521 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="588b97ac4c78bc2adc23c3418b2f1b7f3b8c0146c01099266d14620c807319b3" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.734298 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6" Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.977942 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn"] Oct 14 08:00:03 crc kubenswrapper[5018]: I1014 08:00:03.985323 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-s8jtn"] Oct 14 08:00:04 crc kubenswrapper[5018]: I1014 08:00:04.621352 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7954f62-3572-426b-b442-bccaf1ed9287" path="/var/lib/kubelet/pods/e7954f62-3572-426b-b442-bccaf1ed9287/volumes" Oct 14 08:00:05 crc kubenswrapper[5018]: I1014 08:00:05.073202 5018 scope.go:117] "RemoveContainer" containerID="f26836cf5d46f92d66a1bc414257d46795721b947cf5dbd9e888016bf3d1f3c0" Oct 14 08:00:05 crc kubenswrapper[5018]: I1014 08:00:05.104191 5018 scope.go:117] "RemoveContainer" containerID="8013fe882cdc897e8b3fc977e39e1f617f308b85fde378ff88b9f548a2b146fd" Oct 14 08:00:05 crc kubenswrapper[5018]: I1014 08:00:05.152659 5018 scope.go:117] "RemoveContainer" containerID="74e2c3deb2cba626b48fbfc8d067b4002d9d324c11bace6e3e88e55898d5cbe9" Oct 14 08:00:32 crc kubenswrapper[5018]: I1014 08:00:32.463424 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:00:32 crc kubenswrapper[5018]: I1014 08:00:32.464149 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:02 crc kubenswrapper[5018]: I1014 08:01:02.464081 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:01:02 crc kubenswrapper[5018]: I1014 08:01:02.464988 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.463904 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.464754 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.464821 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.465601 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.465752 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" gracePeriod=600 Oct 14 08:01:32 crc kubenswrapper[5018]: E1014 08:01:32.592117 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.616116 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" exitCode=0 Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.616162 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b"} Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.616193 5018 scope.go:117] "RemoveContainer" containerID="11ed253b45421a2b602d60f0fc5057e0cdfed86029e189e69802a6fda5e0a830" Oct 14 08:01:32 crc kubenswrapper[5018]: I1014 08:01:32.616715 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:01:32 crc kubenswrapper[5018]: E1014 08:01:32.616945 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:01:46 crc kubenswrapper[5018]: I1014 08:01:46.606082 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:01:46 crc kubenswrapper[5018]: E1014 08:01:46.607544 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:01:59 crc kubenswrapper[5018]: I1014 08:01:59.606013 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:01:59 crc kubenswrapper[5018]: E1014 08:01:59.607170 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:02:11 crc kubenswrapper[5018]: I1014 08:02:11.605188 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:02:11 crc kubenswrapper[5018]: E1014 08:02:11.605901 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:02:22 crc kubenswrapper[5018]: I1014 08:02:22.605848 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:02:22 crc kubenswrapper[5018]: E1014 08:02:22.606995 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:02:37 crc kubenswrapper[5018]: I1014 08:02:37.605111 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:02:37 crc kubenswrapper[5018]: E1014 08:02:37.606008 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:02:51 crc kubenswrapper[5018]: I1014 08:02:51.605747 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:02:51 crc kubenswrapper[5018]: E1014 08:02:51.606966 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:02 crc kubenswrapper[5018]: I1014 08:03:02.604898 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:03:02 crc kubenswrapper[5018]: E1014 08:03:02.605996 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:13 crc kubenswrapper[5018]: I1014 08:03:13.605184 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:03:13 crc kubenswrapper[5018]: E1014 08:03:13.606115 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:27 crc kubenswrapper[5018]: I1014 08:03:27.605733 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:03:27 crc kubenswrapper[5018]: E1014 08:03:27.606784 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:40 crc kubenswrapper[5018]: I1014 08:03:40.605678 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:03:40 crc kubenswrapper[5018]: E1014 08:03:40.606757 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:52 crc kubenswrapper[5018]: I1014 08:03:52.605556 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:03:52 crc kubenswrapper[5018]: E1014 08:03:52.606685 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.712681 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:03:56 crc kubenswrapper[5018]: E1014 08:03:56.713619 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" containerName="collect-profiles" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.713671 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" containerName="collect-profiles" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.713912 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" containerName="collect-profiles" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.715760 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.726782 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.850493 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzch5\" (UniqueName: \"kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.851001 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.851230 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.952455 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzch5\" (UniqueName: \"kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.952523 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.952603 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.953085 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:56 crc kubenswrapper[5018]: I1014 08:03:56.953198 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:57 crc kubenswrapper[5018]: I1014 08:03:56.998436 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzch5\" (UniqueName: \"kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5\") pod \"redhat-marketplace-xgttp\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:57 crc kubenswrapper[5018]: I1014 08:03:57.083906 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:03:57 crc kubenswrapper[5018]: I1014 08:03:57.532927 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:03:58 crc kubenswrapper[5018]: I1014 08:03:58.002965 5018 generic.go:334] "Generic (PLEG): container finished" podID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerID="35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd" exitCode=0 Oct 14 08:03:58 crc kubenswrapper[5018]: I1014 08:03:58.003003 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerDied","Data":"35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd"} Oct 14 08:03:58 crc kubenswrapper[5018]: I1014 08:03:58.003033 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerStarted","Data":"05c8001629754803be639e80ffd7d2d68aeffd66ce2d3bcbd00a4bdd42aa6cf8"} Oct 14 08:03:58 crc kubenswrapper[5018]: I1014 08:03:58.005116 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:03:59 crc kubenswrapper[5018]: I1014 08:03:59.017058 5018 generic.go:334] "Generic (PLEG): container finished" podID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerID="bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2" exitCode=0 Oct 14 08:03:59 crc kubenswrapper[5018]: I1014 08:03:59.017128 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerDied","Data":"bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2"} Oct 14 08:04:01 crc kubenswrapper[5018]: I1014 08:04:01.045181 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerStarted","Data":"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7"} Oct 14 08:04:01 crc kubenswrapper[5018]: I1014 08:04:01.082519 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xgttp" podStartSLOduration=3.228021567 podStartE2EDuration="5.082491998s" podCreationTimestamp="2025-10-14 08:03:56 +0000 UTC" firstStartedPulling="2025-10-14 08:03:58.004762277 +0000 UTC m=+4454.588808924" lastFinishedPulling="2025-10-14 08:03:59.859232688 +0000 UTC m=+4456.443279355" observedRunningTime="2025-10-14 08:04:01.075061486 +0000 UTC m=+4457.659108153" watchObservedRunningTime="2025-10-14 08:04:01.082491998 +0000 UTC m=+4457.666538665" Oct 14 08:04:06 crc kubenswrapper[5018]: I1014 08:04:06.605831 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:04:06 crc kubenswrapper[5018]: E1014 08:04:06.606774 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:04:07 crc kubenswrapper[5018]: I1014 08:04:07.084648 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:07 crc kubenswrapper[5018]: I1014 08:04:07.084785 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:07 crc kubenswrapper[5018]: I1014 08:04:07.131119 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:08 crc kubenswrapper[5018]: I1014 08:04:08.195270 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:09 crc kubenswrapper[5018]: I1014 08:04:09.462359 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.142372 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xgttp" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="registry-server" containerID="cri-o://80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7" gracePeriod=2 Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.670391 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.785316 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzch5\" (UniqueName: \"kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5\") pod \"ae71615f-525a-4da6-8759-fd8eb11e70a5\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.785408 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content\") pod \"ae71615f-525a-4da6-8759-fd8eb11e70a5\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.785539 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities\") pod \"ae71615f-525a-4da6-8759-fd8eb11e70a5\" (UID: \"ae71615f-525a-4da6-8759-fd8eb11e70a5\") " Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.787064 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities" (OuterVolumeSpecName: "utilities") pod "ae71615f-525a-4da6-8759-fd8eb11e70a5" (UID: "ae71615f-525a-4da6-8759-fd8eb11e70a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.793245 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5" (OuterVolumeSpecName: "kube-api-access-kzch5") pod "ae71615f-525a-4da6-8759-fd8eb11e70a5" (UID: "ae71615f-525a-4da6-8759-fd8eb11e70a5"). InnerVolumeSpecName "kube-api-access-kzch5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.805270 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae71615f-525a-4da6-8759-fd8eb11e70a5" (UID: "ae71615f-525a-4da6-8759-fd8eb11e70a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.887680 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.888348 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzch5\" (UniqueName: \"kubernetes.io/projected/ae71615f-525a-4da6-8759-fd8eb11e70a5-kube-api-access-kzch5\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:10 crc kubenswrapper[5018]: I1014 08:04:10.888511 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae71615f-525a-4da6-8759-fd8eb11e70a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.159023 5018 generic.go:334] "Generic (PLEG): container finished" podID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerID="80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7" exitCode=0 Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.159117 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerDied","Data":"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7"} Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.159213 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xgttp" event={"ID":"ae71615f-525a-4da6-8759-fd8eb11e70a5","Type":"ContainerDied","Data":"05c8001629754803be639e80ffd7d2d68aeffd66ce2d3bcbd00a4bdd42aa6cf8"} Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.159250 5018 scope.go:117] "RemoveContainer" containerID="80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.159751 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xgttp" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.192935 5018 scope.go:117] "RemoveContainer" containerID="bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.203261 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.208072 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xgttp"] Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.230823 5018 scope.go:117] "RemoveContainer" containerID="35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.271097 5018 scope.go:117] "RemoveContainer" containerID="80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7" Oct 14 08:04:11 crc kubenswrapper[5018]: E1014 08:04:11.271757 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7\": container with ID starting with 80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7 not found: ID does not exist" containerID="80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.271830 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7"} err="failed to get container status \"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7\": rpc error: code = NotFound desc = could not find container \"80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7\": container with ID starting with 80d5e9373d6afec5c1fc8fa7072476aa757ed649785c60970a6bc3d2628004d7 not found: ID does not exist" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.271873 5018 scope.go:117] "RemoveContainer" containerID="bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2" Oct 14 08:04:11 crc kubenswrapper[5018]: E1014 08:04:11.272323 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2\": container with ID starting with bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2 not found: ID does not exist" containerID="bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.272388 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2"} err="failed to get container status \"bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2\": rpc error: code = NotFound desc = could not find container \"bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2\": container with ID starting with bd5087ff193a55cffee2a1df151aab4c465e1abf1c20909e897ced1d1dc319b2 not found: ID does not exist" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.272418 5018 scope.go:117] "RemoveContainer" containerID="35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd" Oct 14 08:04:11 crc kubenswrapper[5018]: E1014 08:04:11.272832 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd\": container with ID starting with 35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd not found: ID does not exist" containerID="35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd" Oct 14 08:04:11 crc kubenswrapper[5018]: I1014 08:04:11.272891 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd"} err="failed to get container status \"35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd\": rpc error: code = NotFound desc = could not find container \"35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd\": container with ID starting with 35196f07f27ae5913284f2634cd23f92118b5d968c2b9c4d6a33e78e018af7dd not found: ID does not exist" Oct 14 08:04:12 crc kubenswrapper[5018]: I1014 08:04:12.644492 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" path="/var/lib/kubelet/pods/ae71615f-525a-4da6-8759-fd8eb11e70a5/volumes" Oct 14 08:04:18 crc kubenswrapper[5018]: I1014 08:04:18.605093 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:04:18 crc kubenswrapper[5018]: E1014 08:04:18.606277 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:04:32 crc kubenswrapper[5018]: I1014 08:04:32.606346 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:04:32 crc kubenswrapper[5018]: E1014 08:04:32.607822 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:04:44 crc kubenswrapper[5018]: I1014 08:04:44.615363 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:04:44 crc kubenswrapper[5018]: E1014 08:04:44.619184 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:04:55 crc kubenswrapper[5018]: I1014 08:04:55.605523 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:04:55 crc kubenswrapper[5018]: E1014 08:04:55.606548 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:05:10 crc kubenswrapper[5018]: I1014 08:05:10.605017 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:05:10 crc kubenswrapper[5018]: E1014 08:05:10.606368 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:05:23 crc kubenswrapper[5018]: I1014 08:05:23.605873 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:05:23 crc kubenswrapper[5018]: E1014 08:05:23.606699 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:05:38 crc kubenswrapper[5018]: I1014 08:05:38.607039 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:05:38 crc kubenswrapper[5018]: E1014 08:05:38.610787 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:05:52 crc kubenswrapper[5018]: I1014 08:05:52.604665 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:05:52 crc kubenswrapper[5018]: E1014 08:05:52.606992 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:06:03 crc kubenswrapper[5018]: I1014 08:06:03.604901 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:06:03 crc kubenswrapper[5018]: E1014 08:06:03.605514 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:06:14 crc kubenswrapper[5018]: I1014 08:06:14.612596 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:06:14 crc kubenswrapper[5018]: E1014 08:06:14.613887 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:06:28 crc kubenswrapper[5018]: I1014 08:06:28.605591 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:06:28 crc kubenswrapper[5018]: E1014 08:06:28.606910 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.281451 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:33 crc kubenswrapper[5018]: E1014 08:06:33.282329 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="registry-server" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.282353 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="registry-server" Oct 14 08:06:33 crc kubenswrapper[5018]: E1014 08:06:33.282396 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="extract-content" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.282409 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="extract-content" Oct 14 08:06:33 crc kubenswrapper[5018]: E1014 08:06:33.282432 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="extract-utilities" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.282445 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="extract-utilities" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.282767 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae71615f-525a-4da6-8759-fd8eb11e70a5" containerName="registry-server" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.289464 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.313673 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.412200 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.412363 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8zhh\" (UniqueName: \"kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.412524 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.514120 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.514597 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8zhh\" (UniqueName: \"kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.514699 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.514950 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.515615 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.544092 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8zhh\" (UniqueName: \"kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh\") pod \"certified-operators-mt4r4\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:33 crc kubenswrapper[5018]: I1014 08:06:33.628876 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:34 crc kubenswrapper[5018]: I1014 08:06:34.103417 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:34 crc kubenswrapper[5018]: W1014 08:06:34.127959 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod275ca64b_7e0e_4aba_a2c6_12c92b043b1d.slice/crio-8d3eac2d776a1f764e6fb8f66ea7cd059db2571385da252a349db29d329d775d WatchSource:0}: Error finding container 8d3eac2d776a1f764e6fb8f66ea7cd059db2571385da252a349db29d329d775d: Status 404 returned error can't find the container with id 8d3eac2d776a1f764e6fb8f66ea7cd059db2571385da252a349db29d329d775d Oct 14 08:06:34 crc kubenswrapper[5018]: I1014 08:06:34.501083 5018 generic.go:334] "Generic (PLEG): container finished" podID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerID="eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7" exitCode=0 Oct 14 08:06:34 crc kubenswrapper[5018]: I1014 08:06:34.501120 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerDied","Data":"eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7"} Oct 14 08:06:34 crc kubenswrapper[5018]: I1014 08:06:34.501143 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerStarted","Data":"8d3eac2d776a1f764e6fb8f66ea7cd059db2571385da252a349db29d329d775d"} Oct 14 08:06:35 crc kubenswrapper[5018]: I1014 08:06:35.512953 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerStarted","Data":"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394"} Oct 14 08:06:36 crc kubenswrapper[5018]: I1014 08:06:36.528700 5018 generic.go:334] "Generic (PLEG): container finished" podID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerID="ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394" exitCode=0 Oct 14 08:06:36 crc kubenswrapper[5018]: I1014 08:06:36.528770 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerDied","Data":"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394"} Oct 14 08:06:37 crc kubenswrapper[5018]: I1014 08:06:37.543160 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerStarted","Data":"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba"} Oct 14 08:06:37 crc kubenswrapper[5018]: I1014 08:06:37.585388 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mt4r4" podStartSLOduration=2.091179199 podStartE2EDuration="4.585359935s" podCreationTimestamp="2025-10-14 08:06:33 +0000 UTC" firstStartedPulling="2025-10-14 08:06:34.502264161 +0000 UTC m=+4611.086310788" lastFinishedPulling="2025-10-14 08:06:36.996444867 +0000 UTC m=+4613.580491524" observedRunningTime="2025-10-14 08:06:37.577496131 +0000 UTC m=+4614.161542798" watchObservedRunningTime="2025-10-14 08:06:37.585359935 +0000 UTC m=+4614.169406592" Oct 14 08:06:40 crc kubenswrapper[5018]: I1014 08:06:40.606287 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:06:41 crc kubenswrapper[5018]: I1014 08:06:41.582343 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc"} Oct 14 08:06:43 crc kubenswrapper[5018]: I1014 08:06:43.630662 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:43 crc kubenswrapper[5018]: I1014 08:06:43.631062 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:43 crc kubenswrapper[5018]: I1014 08:06:43.691385 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:45 crc kubenswrapper[5018]: I1014 08:06:45.046861 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:45 crc kubenswrapper[5018]: I1014 08:06:45.103505 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:46 crc kubenswrapper[5018]: I1014 08:06:46.629914 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mt4r4" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="registry-server" containerID="cri-o://05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba" gracePeriod=2 Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.074079 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.139574 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content\") pod \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.139706 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities\") pod \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.139849 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8zhh\" (UniqueName: \"kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh\") pod \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\" (UID: \"275ca64b-7e0e-4aba-a2c6-12c92b043b1d\") " Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.140758 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities" (OuterVolumeSpecName: "utilities") pod "275ca64b-7e0e-4aba-a2c6-12c92b043b1d" (UID: "275ca64b-7e0e-4aba-a2c6-12c92b043b1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.145995 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh" (OuterVolumeSpecName: "kube-api-access-x8zhh") pod "275ca64b-7e0e-4aba-a2c6-12c92b043b1d" (UID: "275ca64b-7e0e-4aba-a2c6-12c92b043b1d"). InnerVolumeSpecName "kube-api-access-x8zhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.184218 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "275ca64b-7e0e-4aba-a2c6-12c92b043b1d" (UID: "275ca64b-7e0e-4aba-a2c6-12c92b043b1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.241590 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8zhh\" (UniqueName: \"kubernetes.io/projected/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-kube-api-access-x8zhh\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.241681 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.241701 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275ca64b-7e0e-4aba-a2c6-12c92b043b1d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.644669 5018 generic.go:334] "Generic (PLEG): container finished" podID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerID="05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba" exitCode=0 Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.644733 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerDied","Data":"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba"} Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.644772 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mt4r4" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.644797 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mt4r4" event={"ID":"275ca64b-7e0e-4aba-a2c6-12c92b043b1d","Type":"ContainerDied","Data":"8d3eac2d776a1f764e6fb8f66ea7cd059db2571385da252a349db29d329d775d"} Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.644835 5018 scope.go:117] "RemoveContainer" containerID="05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.701206 5018 scope.go:117] "RemoveContainer" containerID="ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.710456 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.716810 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mt4r4"] Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.729786 5018 scope.go:117] "RemoveContainer" containerID="eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.771785 5018 scope.go:117] "RemoveContainer" containerID="05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba" Oct 14 08:06:47 crc kubenswrapper[5018]: E1014 08:06:47.772543 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba\": container with ID starting with 05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba not found: ID does not exist" containerID="05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.772591 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba"} err="failed to get container status \"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba\": rpc error: code = NotFound desc = could not find container \"05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba\": container with ID starting with 05dbbe9de0eea55039656f18ee61336dc5c007888e536d3132cfcbb14948cbba not found: ID does not exist" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.772661 5018 scope.go:117] "RemoveContainer" containerID="ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394" Oct 14 08:06:47 crc kubenswrapper[5018]: E1014 08:06:47.773302 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394\": container with ID starting with ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394 not found: ID does not exist" containerID="ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.773331 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394"} err="failed to get container status \"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394\": rpc error: code = NotFound desc = could not find container \"ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394\": container with ID starting with ca7d6fd9b43107f894c8fafe2ebffda1494a98e43f79adee4ebd3e01a9bbb394 not found: ID does not exist" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.773350 5018 scope.go:117] "RemoveContainer" containerID="eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7" Oct 14 08:06:47 crc kubenswrapper[5018]: E1014 08:06:47.773882 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7\": container with ID starting with eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7 not found: ID does not exist" containerID="eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7" Oct 14 08:06:47 crc kubenswrapper[5018]: I1014 08:06:47.773927 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7"} err="failed to get container status \"eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7\": rpc error: code = NotFound desc = could not find container \"eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7\": container with ID starting with eb6d5201d4ea4965f9d5021bb1c00e2608cf0aa742a24ec05702ffee09f52df7 not found: ID does not exist" Oct 14 08:06:48 crc kubenswrapper[5018]: I1014 08:06:48.622674 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" path="/var/lib/kubelet/pods/275ca64b-7e0e-4aba-a2c6-12c92b043b1d/volumes" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.937948 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:06:59 crc kubenswrapper[5018]: E1014 08:06:59.938981 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="extract-content" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.938999 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="extract-content" Oct 14 08:06:59 crc kubenswrapper[5018]: E1014 08:06:59.939019 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="registry-server" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.939028 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="registry-server" Oct 14 08:06:59 crc kubenswrapper[5018]: E1014 08:06:59.939044 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="extract-utilities" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.939052 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="extract-utilities" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.939247 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="275ca64b-7e0e-4aba-a2c6-12c92b043b1d" containerName="registry-server" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.940538 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:06:59 crc kubenswrapper[5018]: I1014 08:06:59.959251 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.045355 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfm4b\" (UniqueName: \"kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.045464 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.045501 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.146848 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.147488 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.147565 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.147883 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.148082 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfm4b\" (UniqueName: \"kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.176999 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfm4b\" (UniqueName: \"kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b\") pod \"redhat-operators-ffsws\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.272073 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.713961 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:07:00 crc kubenswrapper[5018]: I1014 08:07:00.770692 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerStarted","Data":"8bd41f104dca2cf439c54ec0528b82a55b5d49e863fa472b11fd57b20bc550d9"} Oct 14 08:07:01 crc kubenswrapper[5018]: I1014 08:07:01.782506 5018 generic.go:334] "Generic (PLEG): container finished" podID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerID="62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4" exitCode=0 Oct 14 08:07:01 crc kubenswrapper[5018]: I1014 08:07:01.782585 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerDied","Data":"62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4"} Oct 14 08:07:03 crc kubenswrapper[5018]: I1014 08:07:03.797891 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerStarted","Data":"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d"} Oct 14 08:07:04 crc kubenswrapper[5018]: I1014 08:07:04.810418 5018 generic.go:334] "Generic (PLEG): container finished" podID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerID="116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d" exitCode=0 Oct 14 08:07:04 crc kubenswrapper[5018]: I1014 08:07:04.811069 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerDied","Data":"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d"} Oct 14 08:07:06 crc kubenswrapper[5018]: I1014 08:07:06.832355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerStarted","Data":"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf"} Oct 14 08:07:06 crc kubenswrapper[5018]: I1014 08:07:06.862087 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ffsws" podStartSLOduration=3.995444467 podStartE2EDuration="7.862061776s" podCreationTimestamp="2025-10-14 08:06:59 +0000 UTC" firstStartedPulling="2025-10-14 08:07:01.784599461 +0000 UTC m=+4638.368646088" lastFinishedPulling="2025-10-14 08:07:05.65121673 +0000 UTC m=+4642.235263397" observedRunningTime="2025-10-14 08:07:06.859808952 +0000 UTC m=+4643.443855619" watchObservedRunningTime="2025-10-14 08:07:06.862061776 +0000 UTC m=+4643.446108443" Oct 14 08:07:10 crc kubenswrapper[5018]: I1014 08:07:10.273062 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:10 crc kubenswrapper[5018]: I1014 08:07:10.273449 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:11 crc kubenswrapper[5018]: I1014 08:07:11.322117 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ffsws" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="registry-server" probeResult="failure" output=< Oct 14 08:07:11 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 08:07:11 crc kubenswrapper[5018]: > Oct 14 08:07:20 crc kubenswrapper[5018]: I1014 08:07:20.357381 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:20 crc kubenswrapper[5018]: I1014 08:07:20.404524 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:20 crc kubenswrapper[5018]: I1014 08:07:20.598215 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:07:21 crc kubenswrapper[5018]: I1014 08:07:21.972550 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ffsws" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="registry-server" containerID="cri-o://74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf" gracePeriod=2 Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.348240 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.500184 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content\") pod \"070517f7-0456-42a0-aeeb-ac0d843d01fd\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.500258 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities\") pod \"070517f7-0456-42a0-aeeb-ac0d843d01fd\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.500393 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfm4b\" (UniqueName: \"kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b\") pod \"070517f7-0456-42a0-aeeb-ac0d843d01fd\" (UID: \"070517f7-0456-42a0-aeeb-ac0d843d01fd\") " Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.501891 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities" (OuterVolumeSpecName: "utilities") pod "070517f7-0456-42a0-aeeb-ac0d843d01fd" (UID: "070517f7-0456-42a0-aeeb-ac0d843d01fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.508886 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b" (OuterVolumeSpecName: "kube-api-access-xfm4b") pod "070517f7-0456-42a0-aeeb-ac0d843d01fd" (UID: "070517f7-0456-42a0-aeeb-ac0d843d01fd"). InnerVolumeSpecName "kube-api-access-xfm4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.587059 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "070517f7-0456-42a0-aeeb-ac0d843d01fd" (UID: "070517f7-0456-42a0-aeeb-ac0d843d01fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.602473 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfm4b\" (UniqueName: \"kubernetes.io/projected/070517f7-0456-42a0-aeeb-ac0d843d01fd-kube-api-access-xfm4b\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.602523 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.602536 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/070517f7-0456-42a0-aeeb-ac0d843d01fd-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.983864 5018 generic.go:334] "Generic (PLEG): container finished" podID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerID="74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf" exitCode=0 Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.983912 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerDied","Data":"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf"} Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.983946 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ffsws" event={"ID":"070517f7-0456-42a0-aeeb-ac0d843d01fd","Type":"ContainerDied","Data":"8bd41f104dca2cf439c54ec0528b82a55b5d49e863fa472b11fd57b20bc550d9"} Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.983946 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ffsws" Oct 14 08:07:22 crc kubenswrapper[5018]: I1014 08:07:22.983966 5018 scope.go:117] "RemoveContainer" containerID="74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.010270 5018 scope.go:117] "RemoveContainer" containerID="116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.014101 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.020790 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ffsws"] Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.034158 5018 scope.go:117] "RemoveContainer" containerID="62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.060714 5018 scope.go:117] "RemoveContainer" containerID="74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf" Oct 14 08:07:23 crc kubenswrapper[5018]: E1014 08:07:23.061412 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf\": container with ID starting with 74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf not found: ID does not exist" containerID="74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.061451 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf"} err="failed to get container status \"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf\": rpc error: code = NotFound desc = could not find container \"74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf\": container with ID starting with 74405b8e7a6b4ea7a7db63989a0709854b26e5f184d0477b1e63e819ad95a1bf not found: ID does not exist" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.061476 5018 scope.go:117] "RemoveContainer" containerID="116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d" Oct 14 08:07:23 crc kubenswrapper[5018]: E1014 08:07:23.062475 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d\": container with ID starting with 116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d not found: ID does not exist" containerID="116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.062534 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d"} err="failed to get container status \"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d\": rpc error: code = NotFound desc = could not find container \"116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d\": container with ID starting with 116f49da0f8f76bbe88d95f66d3ddd32d6e5b47e0844c57432d1f00ce323567d not found: ID does not exist" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.062576 5018 scope.go:117] "RemoveContainer" containerID="62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4" Oct 14 08:07:23 crc kubenswrapper[5018]: E1014 08:07:23.063224 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4\": container with ID starting with 62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4 not found: ID does not exist" containerID="62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4" Oct 14 08:07:23 crc kubenswrapper[5018]: I1014 08:07:23.063263 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4"} err="failed to get container status \"62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4\": rpc error: code = NotFound desc = could not find container \"62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4\": container with ID starting with 62ce7138c303dcdcd9116dafa804616a348a35315e5cf213ad6d43305c771cd4 not found: ID does not exist" Oct 14 08:07:24 crc kubenswrapper[5018]: I1014 08:07:24.620281 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" path="/var/lib/kubelet/pods/070517f7-0456-42a0-aeeb-ac0d843d01fd/volumes" Oct 14 08:09:02 crc kubenswrapper[5018]: I1014 08:09:02.463235 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:09:02 crc kubenswrapper[5018]: I1014 08:09:02.464104 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:09:32 crc kubenswrapper[5018]: I1014 08:09:32.464223 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:09:32 crc kubenswrapper[5018]: I1014 08:09:32.464849 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:10:02 crc kubenswrapper[5018]: I1014 08:10:02.463805 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:10:02 crc kubenswrapper[5018]: I1014 08:10:02.464542 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:10:02 crc kubenswrapper[5018]: I1014 08:10:02.464594 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:10:02 crc kubenswrapper[5018]: I1014 08:10:02.465328 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:10:02 crc kubenswrapper[5018]: I1014 08:10:02.465397 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc" gracePeriod=600 Oct 14 08:10:03 crc kubenswrapper[5018]: I1014 08:10:03.425365 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc" exitCode=0 Oct 14 08:10:03 crc kubenswrapper[5018]: I1014 08:10:03.425436 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc"} Oct 14 08:10:03 crc kubenswrapper[5018]: I1014 08:10:03.425762 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f"} Oct 14 08:10:03 crc kubenswrapper[5018]: I1014 08:10:03.425790 5018 scope.go:117] "RemoveContainer" containerID="68a3d76dd23f10d6524fcd82e7e21e1ab6f93c5723b2897b4d4b29fa9ff0935b" Oct 14 08:12:02 crc kubenswrapper[5018]: I1014 08:12:02.463518 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:12:02 crc kubenswrapper[5018]: I1014 08:12:02.464226 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:12:32 crc kubenswrapper[5018]: I1014 08:12:32.463833 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:12:32 crc kubenswrapper[5018]: I1014 08:12:32.464524 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:13:02 crc kubenswrapper[5018]: I1014 08:13:02.464135 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:13:02 crc kubenswrapper[5018]: I1014 08:13:02.464795 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:13:02 crc kubenswrapper[5018]: I1014 08:13:02.464853 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:13:02 crc kubenswrapper[5018]: I1014 08:13:02.465560 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:13:02 crc kubenswrapper[5018]: I1014 08:13:02.465699 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" gracePeriod=600 Oct 14 08:13:02 crc kubenswrapper[5018]: E1014 08:13:02.588605 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:13:03 crc kubenswrapper[5018]: I1014 08:13:03.088950 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" exitCode=0 Oct 14 08:13:03 crc kubenswrapper[5018]: I1014 08:13:03.089040 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f"} Oct 14 08:13:03 crc kubenswrapper[5018]: I1014 08:13:03.089093 5018 scope.go:117] "RemoveContainer" containerID="5253551c8f7a644192e8163ecf557fa59198d8b6a253adf09267c6f4744970fc" Oct 14 08:13:03 crc kubenswrapper[5018]: I1014 08:13:03.089912 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:13:03 crc kubenswrapper[5018]: E1014 08:13:03.090350 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.432325 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:07 crc kubenswrapper[5018]: E1014 08:13:07.433945 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="extract-utilities" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.433968 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="extract-utilities" Oct 14 08:13:07 crc kubenswrapper[5018]: E1014 08:13:07.434027 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="registry-server" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.434061 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="registry-server" Oct 14 08:13:07 crc kubenswrapper[5018]: E1014 08:13:07.434149 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="extract-content" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.434199 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="extract-content" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.436150 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="070517f7-0456-42a0-aeeb-ac0d843d01fd" containerName="registry-server" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.440029 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.441758 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.539480 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5jg6\" (UniqueName: \"kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.539906 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.540104 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.641847 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.641898 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.641981 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5jg6\" (UniqueName: \"kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.642494 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:07 crc kubenswrapper[5018]: I1014 08:13:07.642537 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:08 crc kubenswrapper[5018]: I1014 08:13:08.079980 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5jg6\" (UniqueName: \"kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6\") pod \"community-operators-vgm6d\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:08 crc kubenswrapper[5018]: I1014 08:13:08.379093 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:08 crc kubenswrapper[5018]: I1014 08:13:08.621800 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:09 crc kubenswrapper[5018]: I1014 08:13:09.145315 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerID="531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de" exitCode=0 Oct 14 08:13:09 crc kubenswrapper[5018]: I1014 08:13:09.145354 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerDied","Data":"531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de"} Oct 14 08:13:09 crc kubenswrapper[5018]: I1014 08:13:09.145377 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerStarted","Data":"1b4374f64ceec3b1ff68c63a248ad6b5e43c1e8d1c73fc238edda39964c933f5"} Oct 14 08:13:09 crc kubenswrapper[5018]: I1014 08:13:09.147726 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:13:10 crc kubenswrapper[5018]: I1014 08:13:10.158075 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerStarted","Data":"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2"} Oct 14 08:13:11 crc kubenswrapper[5018]: I1014 08:13:11.167527 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerID="4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2" exitCode=0 Oct 14 08:13:11 crc kubenswrapper[5018]: I1014 08:13:11.167569 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerDied","Data":"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2"} Oct 14 08:13:12 crc kubenswrapper[5018]: I1014 08:13:12.184945 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerStarted","Data":"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf"} Oct 14 08:13:12 crc kubenswrapper[5018]: I1014 08:13:12.223149 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vgm6d" podStartSLOduration=2.436261249 podStartE2EDuration="5.223124841s" podCreationTimestamp="2025-10-14 08:13:07 +0000 UTC" firstStartedPulling="2025-10-14 08:13:09.147500587 +0000 UTC m=+5005.731547214" lastFinishedPulling="2025-10-14 08:13:11.934364139 +0000 UTC m=+5008.518410806" observedRunningTime="2025-10-14 08:13:12.214215899 +0000 UTC m=+5008.798262566" watchObservedRunningTime="2025-10-14 08:13:12.223124841 +0000 UTC m=+5008.807171478" Oct 14 08:13:13 crc kubenswrapper[5018]: I1014 08:13:13.604944 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:13:13 crc kubenswrapper[5018]: E1014 08:13:13.605424 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:13:18 crc kubenswrapper[5018]: I1014 08:13:18.379350 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:18 crc kubenswrapper[5018]: I1014 08:13:18.379880 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:18 crc kubenswrapper[5018]: I1014 08:13:18.424490 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:19 crc kubenswrapper[5018]: I1014 08:13:19.317094 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.423543 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.425356 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vgm6d" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="registry-server" containerID="cri-o://02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf" gracePeriod=2 Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.902190 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.960956 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content\") pod \"cc57628d-2ac2-408e-a646-0fe1b2851da8\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.961186 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities\") pod \"cc57628d-2ac2-408e-a646-0fe1b2851da8\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.961232 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5jg6\" (UniqueName: \"kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6\") pod \"cc57628d-2ac2-408e-a646-0fe1b2851da8\" (UID: \"cc57628d-2ac2-408e-a646-0fe1b2851da8\") " Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.962204 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities" (OuterVolumeSpecName: "utilities") pod "cc57628d-2ac2-408e-a646-0fe1b2851da8" (UID: "cc57628d-2ac2-408e-a646-0fe1b2851da8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.962889 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:13:21 crc kubenswrapper[5018]: I1014 08:13:21.969868 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6" (OuterVolumeSpecName: "kube-api-access-d5jg6") pod "cc57628d-2ac2-408e-a646-0fe1b2851da8" (UID: "cc57628d-2ac2-408e-a646-0fe1b2851da8"). InnerVolumeSpecName "kube-api-access-d5jg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.024824 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc57628d-2ac2-408e-a646-0fe1b2851da8" (UID: "cc57628d-2ac2-408e-a646-0fe1b2851da8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.064661 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5jg6\" (UniqueName: \"kubernetes.io/projected/cc57628d-2ac2-408e-a646-0fe1b2851da8-kube-api-access-d5jg6\") on node \"crc\" DevicePath \"\"" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.064697 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc57628d-2ac2-408e-a646-0fe1b2851da8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.286518 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerID="02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf" exitCode=0 Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.286585 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerDied","Data":"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf"} Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.286667 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgm6d" event={"ID":"cc57628d-2ac2-408e-a646-0fe1b2851da8","Type":"ContainerDied","Data":"1b4374f64ceec3b1ff68c63a248ad6b5e43c1e8d1c73fc238edda39964c933f5"} Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.286672 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgm6d" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.286699 5018 scope.go:117] "RemoveContainer" containerID="02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.312742 5018 scope.go:117] "RemoveContainer" containerID="4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.330690 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.336657 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vgm6d"] Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.359420 5018 scope.go:117] "RemoveContainer" containerID="531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.376990 5018 scope.go:117] "RemoveContainer" containerID="02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf" Oct 14 08:13:22 crc kubenswrapper[5018]: E1014 08:13:22.378749 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf\": container with ID starting with 02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf not found: ID does not exist" containerID="02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.378791 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf"} err="failed to get container status \"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf\": rpc error: code = NotFound desc = could not find container \"02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf\": container with ID starting with 02654635576d7e8d6543b0135302f714d67849b9ecbe3d114cba7946b7e1a2bf not found: ID does not exist" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.378831 5018 scope.go:117] "RemoveContainer" containerID="4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2" Oct 14 08:13:22 crc kubenswrapper[5018]: E1014 08:13:22.379345 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2\": container with ID starting with 4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2 not found: ID does not exist" containerID="4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.379370 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2"} err="failed to get container status \"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2\": rpc error: code = NotFound desc = could not find container \"4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2\": container with ID starting with 4d0323b5b97b420b0bf5d048dc4ea96ba8de46b60a6c7fd4ed0983f8f73533a2 not found: ID does not exist" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.379383 5018 scope.go:117] "RemoveContainer" containerID="531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de" Oct 14 08:13:22 crc kubenswrapper[5018]: E1014 08:13:22.379874 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de\": container with ID starting with 531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de not found: ID does not exist" containerID="531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.379947 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de"} err="failed to get container status \"531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de\": rpc error: code = NotFound desc = could not find container \"531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de\": container with ID starting with 531dd77384fc4f5a7483489bd79d69a6c2bc6da3fa128029f43e332ac487f3de not found: ID does not exist" Oct 14 08:13:22 crc kubenswrapper[5018]: I1014 08:13:22.621134 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" path="/var/lib/kubelet/pods/cc57628d-2ac2-408e-a646-0fe1b2851da8/volumes" Oct 14 08:13:26 crc kubenswrapper[5018]: I1014 08:13:26.605265 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:13:26 crc kubenswrapper[5018]: E1014 08:13:26.605740 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:13:40 crc kubenswrapper[5018]: I1014 08:13:40.604886 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:13:40 crc kubenswrapper[5018]: E1014 08:13:40.605843 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:13:55 crc kubenswrapper[5018]: I1014 08:13:55.605420 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:13:55 crc kubenswrapper[5018]: E1014 08:13:55.606454 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:14:06 crc kubenswrapper[5018]: I1014 08:14:06.606434 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:14:06 crc kubenswrapper[5018]: E1014 08:14:06.607462 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:14:20 crc kubenswrapper[5018]: I1014 08:14:20.608267 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:14:20 crc kubenswrapper[5018]: E1014 08:14:20.609278 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:14:32 crc kubenswrapper[5018]: I1014 08:14:32.605596 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:14:32 crc kubenswrapper[5018]: E1014 08:14:32.606492 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:14:46 crc kubenswrapper[5018]: I1014 08:14:46.606003 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:14:46 crc kubenswrapper[5018]: E1014 08:14:46.607230 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.685469 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:14:56 crc kubenswrapper[5018]: E1014 08:14:56.686658 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="extract-utilities" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.686686 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="extract-utilities" Oct 14 08:14:56 crc kubenswrapper[5018]: E1014 08:14:56.686741 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="registry-server" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.686755 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="registry-server" Oct 14 08:14:56 crc kubenswrapper[5018]: E1014 08:14:56.686773 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="extract-content" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.686787 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="extract-content" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.687102 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc57628d-2ac2-408e-a646-0fe1b2851da8" containerName="registry-server" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.689123 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.699097 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.787765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4nv8\" (UniqueName: \"kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.787883 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.787917 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.889027 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4nv8\" (UniqueName: \"kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.889125 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.889158 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.889877 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.889976 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:56 crc kubenswrapper[5018]: I1014 08:14:56.914451 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4nv8\" (UniqueName: \"kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8\") pod \"redhat-marketplace-fgx6m\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:57 crc kubenswrapper[5018]: I1014 08:14:57.006835 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:14:57 crc kubenswrapper[5018]: I1014 08:14:57.498514 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:14:58 crc kubenswrapper[5018]: I1014 08:14:58.386914 5018 generic.go:334] "Generic (PLEG): container finished" podID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerID="936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3" exitCode=0 Oct 14 08:14:58 crc kubenswrapper[5018]: I1014 08:14:58.387229 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerDied","Data":"936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3"} Oct 14 08:14:58 crc kubenswrapper[5018]: I1014 08:14:58.387540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerStarted","Data":"3111f296c43b53530f7b305372e4a3c5a03102fa60a1e4d4e3a5dca1eeb1f6c7"} Oct 14 08:14:59 crc kubenswrapper[5018]: I1014 08:14:59.397291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerStarted","Data":"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7"} Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.161892 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv"] Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.162808 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.166360 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.179577 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.181054 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv"] Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.287151 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.287254 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.287316 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n72pt\" (UniqueName: \"kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.388208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n72pt\" (UniqueName: \"kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.388333 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.388378 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.390003 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.400599 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.412459 5018 generic.go:334] "Generic (PLEG): container finished" podID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerID="76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7" exitCode=0 Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.412857 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerDied","Data":"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7"} Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.415868 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n72pt\" (UniqueName: \"kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt\") pod \"collect-profiles-29340495-jr2lv\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.494955 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:00 crc kubenswrapper[5018]: I1014 08:15:00.973602 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv"] Oct 14 08:15:00 crc kubenswrapper[5018]: W1014 08:15:00.978935 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5357a32_c284_41a6_b927_b36cc1ee6509.slice/crio-16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d WatchSource:0}: Error finding container 16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d: Status 404 returned error can't find the container with id 16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.420410 5018 generic.go:334] "Generic (PLEG): container finished" podID="f5357a32-c284-41a6-b927-b36cc1ee6509" containerID="76ccaf93fe7fdd9fe41953b857e47baabdb375d2fba0a30c8256d88e04add86c" exitCode=0 Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.420513 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" event={"ID":"f5357a32-c284-41a6-b927-b36cc1ee6509","Type":"ContainerDied","Data":"76ccaf93fe7fdd9fe41953b857e47baabdb375d2fba0a30c8256d88e04add86c"} Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.420945 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" event={"ID":"f5357a32-c284-41a6-b927-b36cc1ee6509","Type":"ContainerStarted","Data":"16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d"} Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.423153 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerStarted","Data":"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5"} Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.457323 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fgx6m" podStartSLOduration=2.786946713 podStartE2EDuration="5.457303433s" podCreationTimestamp="2025-10-14 08:14:56 +0000 UTC" firstStartedPulling="2025-10-14 08:14:58.388990966 +0000 UTC m=+5114.973037593" lastFinishedPulling="2025-10-14 08:15:01.059347686 +0000 UTC m=+5117.643394313" observedRunningTime="2025-10-14 08:15:01.454414811 +0000 UTC m=+5118.038461438" watchObservedRunningTime="2025-10-14 08:15:01.457303433 +0000 UTC m=+5118.041350080" Oct 14 08:15:01 crc kubenswrapper[5018]: I1014 08:15:01.604848 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:15:01 crc kubenswrapper[5018]: E1014 08:15:01.605145 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.778644 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.927966 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume\") pod \"f5357a32-c284-41a6-b927-b36cc1ee6509\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.928100 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume\") pod \"f5357a32-c284-41a6-b927-b36cc1ee6509\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.928166 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n72pt\" (UniqueName: \"kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt\") pod \"f5357a32-c284-41a6-b927-b36cc1ee6509\" (UID: \"f5357a32-c284-41a6-b927-b36cc1ee6509\") " Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.929413 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume" (OuterVolumeSpecName: "config-volume") pod "f5357a32-c284-41a6-b927-b36cc1ee6509" (UID: "f5357a32-c284-41a6-b927-b36cc1ee6509"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.934340 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f5357a32-c284-41a6-b927-b36cc1ee6509" (UID: "f5357a32-c284-41a6-b927-b36cc1ee6509"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:02 crc kubenswrapper[5018]: I1014 08:15:02.936056 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt" (OuterVolumeSpecName: "kube-api-access-n72pt") pod "f5357a32-c284-41a6-b927-b36cc1ee6509" (UID: "f5357a32-c284-41a6-b927-b36cc1ee6509"). InnerVolumeSpecName "kube-api-access-n72pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.030759 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5357a32-c284-41a6-b927-b36cc1ee6509-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.030818 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n72pt\" (UniqueName: \"kubernetes.io/projected/f5357a32-c284-41a6-b927-b36cc1ee6509-kube-api-access-n72pt\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.030838 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5357a32-c284-41a6-b927-b36cc1ee6509-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.443758 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" event={"ID":"f5357a32-c284-41a6-b927-b36cc1ee6509","Type":"ContainerDied","Data":"16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d"} Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.443809 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.443805 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16cfd4df49afd17966f973bcef25b683ae0b8cbb5b3644ad0f18d1895c3ce48d" Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.864922 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg"] Oct 14 08:15:03 crc kubenswrapper[5018]: I1014 08:15:03.869428 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-7sdsg"] Oct 14 08:15:04 crc kubenswrapper[5018]: I1014 08:15:04.622580 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6e546e7-fc9e-4719-9b9f-336487c6e19f" path="/var/lib/kubelet/pods/c6e546e7-fc9e-4719-9b9f-336487c6e19f/volumes" Oct 14 08:15:05 crc kubenswrapper[5018]: I1014 08:15:05.566784 5018 scope.go:117] "RemoveContainer" containerID="b23306eb3f6d47deb83cd501e2659455c356aede3644c92e83583882f777e7ad" Oct 14 08:15:07 crc kubenswrapper[5018]: I1014 08:15:07.007148 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:07 crc kubenswrapper[5018]: I1014 08:15:07.007875 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:07 crc kubenswrapper[5018]: I1014 08:15:07.086130 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:07 crc kubenswrapper[5018]: I1014 08:15:07.540462 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:07 crc kubenswrapper[5018]: I1014 08:15:07.605612 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:15:09 crc kubenswrapper[5018]: I1014 08:15:09.500723 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fgx6m" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="registry-server" containerID="cri-o://e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5" gracePeriod=2 Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.010361 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.061144 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content\") pod \"7a729b89-b820-40e0-b125-8e6e78c6dee4\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.061686 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4nv8\" (UniqueName: \"kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8\") pod \"7a729b89-b820-40e0-b125-8e6e78c6dee4\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.061893 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities\") pod \"7a729b89-b820-40e0-b125-8e6e78c6dee4\" (UID: \"7a729b89-b820-40e0-b125-8e6e78c6dee4\") " Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.063039 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities" (OuterVolumeSpecName: "utilities") pod "7a729b89-b820-40e0-b125-8e6e78c6dee4" (UID: "7a729b89-b820-40e0-b125-8e6e78c6dee4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.068588 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8" (OuterVolumeSpecName: "kube-api-access-z4nv8") pod "7a729b89-b820-40e0-b125-8e6e78c6dee4" (UID: "7a729b89-b820-40e0-b125-8e6e78c6dee4"). InnerVolumeSpecName "kube-api-access-z4nv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.084843 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a729b89-b820-40e0-b125-8e6e78c6dee4" (UID: "7a729b89-b820-40e0-b125-8e6e78c6dee4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.163001 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.163065 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a729b89-b820-40e0-b125-8e6e78c6dee4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.163085 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4nv8\" (UniqueName: \"kubernetes.io/projected/7a729b89-b820-40e0-b125-8e6e78c6dee4-kube-api-access-z4nv8\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.514633 5018 generic.go:334] "Generic (PLEG): container finished" podID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerID="e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5" exitCode=0 Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.514657 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerDied","Data":"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5"} Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.514710 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgx6m" event={"ID":"7a729b89-b820-40e0-b125-8e6e78c6dee4","Type":"ContainerDied","Data":"3111f296c43b53530f7b305372e4a3c5a03102fa60a1e4d4e3a5dca1eeb1f6c7"} Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.514729 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgx6m" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.514738 5018 scope.go:117] "RemoveContainer" containerID="e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.558932 5018 scope.go:117] "RemoveContainer" containerID="76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.579666 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.593543 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgx6m"] Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.612448 5018 scope.go:117] "RemoveContainer" containerID="936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.622210 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" path="/var/lib/kubelet/pods/7a729b89-b820-40e0-b125-8e6e78c6dee4/volumes" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.640220 5018 scope.go:117] "RemoveContainer" containerID="e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5" Oct 14 08:15:10 crc kubenswrapper[5018]: E1014 08:15:10.640782 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5\": container with ID starting with e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5 not found: ID does not exist" containerID="e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.640840 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5"} err="failed to get container status \"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5\": rpc error: code = NotFound desc = could not find container \"e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5\": container with ID starting with e559991febc21553f97e21a86e21aa077f680ad75d32711aa8d929d9ea264ed5 not found: ID does not exist" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.640875 5018 scope.go:117] "RemoveContainer" containerID="76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7" Oct 14 08:15:10 crc kubenswrapper[5018]: E1014 08:15:10.641360 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7\": container with ID starting with 76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7 not found: ID does not exist" containerID="76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.641405 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7"} err="failed to get container status \"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7\": rpc error: code = NotFound desc = could not find container \"76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7\": container with ID starting with 76dd483d66a4bb57925d63a44858ed8cb5874b34a26938a79a96f55c939772d7 not found: ID does not exist" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.641433 5018 scope.go:117] "RemoveContainer" containerID="936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3" Oct 14 08:15:10 crc kubenswrapper[5018]: E1014 08:15:10.641888 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3\": container with ID starting with 936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3 not found: ID does not exist" containerID="936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3" Oct 14 08:15:10 crc kubenswrapper[5018]: I1014 08:15:10.641953 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3"} err="failed to get container status \"936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3\": rpc error: code = NotFound desc = could not find container \"936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3\": container with ID starting with 936baa031710b92a86f844b9cc97c4a33a8a1d9d764d3bbf0d05b6415c6d39e3 not found: ID does not exist" Oct 14 08:15:16 crc kubenswrapper[5018]: I1014 08:15:16.606464 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:15:16 crc kubenswrapper[5018]: E1014 08:15:16.607600 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:15:31 crc kubenswrapper[5018]: I1014 08:15:31.605540 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:15:31 crc kubenswrapper[5018]: E1014 08:15:31.606324 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:15:44 crc kubenswrapper[5018]: I1014 08:15:44.615333 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:15:44 crc kubenswrapper[5018]: E1014 08:15:44.617012 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:15:59 crc kubenswrapper[5018]: I1014 08:15:59.605452 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:15:59 crc kubenswrapper[5018]: E1014 08:15:59.606335 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:16:13 crc kubenswrapper[5018]: I1014 08:16:13.605858 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:16:13 crc kubenswrapper[5018]: E1014 08:16:13.606677 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:16:25 crc kubenswrapper[5018]: I1014 08:16:25.605504 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:16:25 crc kubenswrapper[5018]: E1014 08:16:25.606678 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:16:39 crc kubenswrapper[5018]: I1014 08:16:39.606235 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:16:39 crc kubenswrapper[5018]: E1014 08:16:39.606947 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:16:54 crc kubenswrapper[5018]: I1014 08:16:54.608249 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:16:54 crc kubenswrapper[5018]: E1014 08:16:54.608938 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:17:07 crc kubenswrapper[5018]: I1014 08:17:07.605865 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:17:07 crc kubenswrapper[5018]: E1014 08:17:07.607239 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.660024 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:09 crc kubenswrapper[5018]: E1014 08:17:09.660994 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5357a32-c284-41a6-b927-b36cc1ee6509" containerName="collect-profiles" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661017 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5357a32-c284-41a6-b927-b36cc1ee6509" containerName="collect-profiles" Oct 14 08:17:09 crc kubenswrapper[5018]: E1014 08:17:09.661042 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="extract-content" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661055 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="extract-content" Oct 14 08:17:09 crc kubenswrapper[5018]: E1014 08:17:09.661092 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="registry-server" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661105 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="registry-server" Oct 14 08:17:09 crc kubenswrapper[5018]: E1014 08:17:09.661136 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="extract-utilities" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661150 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="extract-utilities" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661446 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5357a32-c284-41a6-b927-b36cc1ee6509" containerName="collect-profiles" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.661485 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a729b89-b820-40e0-b125-8e6e78c6dee4" containerName="registry-server" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.663600 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.684399 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.814322 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.814740 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm8hb\" (UniqueName: \"kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.814985 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.916189 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.916569 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.916717 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.916920 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm8hb\" (UniqueName: \"kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.916978 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.936496 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm8hb\" (UniqueName: \"kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb\") pod \"certified-operators-cd6jd\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:09 crc kubenswrapper[5018]: I1014 08:17:09.993585 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:10 crc kubenswrapper[5018]: I1014 08:17:10.482009 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:10 crc kubenswrapper[5018]: I1014 08:17:10.715509 5018 generic.go:334] "Generic (PLEG): container finished" podID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerID="81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5" exitCode=0 Oct 14 08:17:10 crc kubenswrapper[5018]: I1014 08:17:10.715582 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerDied","Data":"81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5"} Oct 14 08:17:10 crc kubenswrapper[5018]: I1014 08:17:10.717363 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerStarted","Data":"2874104d183fd3c22426266a7563af96adaaae287d594905d8e6b0ca48e7a831"} Oct 14 08:17:11 crc kubenswrapper[5018]: I1014 08:17:11.739381 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerStarted","Data":"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b"} Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.433658 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.436354 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.450776 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.560525 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnbbf\" (UniqueName: \"kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.560917 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.561014 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.663712 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.663776 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.664010 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnbbf\" (UniqueName: \"kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.665123 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.665535 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.692872 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnbbf\" (UniqueName: \"kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf\") pod \"redhat-operators-th7t9\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.751374 5018 generic.go:334] "Generic (PLEG): container finished" podID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerID="9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b" exitCode=0 Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.751448 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerDied","Data":"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b"} Oct 14 08:17:12 crc kubenswrapper[5018]: I1014 08:17:12.799367 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.064509 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.760490 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerStarted","Data":"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9"} Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.763844 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerID="5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9" exitCode=0 Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.763895 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerDied","Data":"5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9"} Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.763918 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerStarted","Data":"086110c0400db100c9537c324f4979a5c9fd7a53a39abfacdbd55ca2a644937e"} Oct 14 08:17:13 crc kubenswrapper[5018]: I1014 08:17:13.801970 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cd6jd" podStartSLOduration=2.039237367 podStartE2EDuration="4.801947084s" podCreationTimestamp="2025-10-14 08:17:09 +0000 UTC" firstStartedPulling="2025-10-14 08:17:10.719863037 +0000 UTC m=+5247.303909704" lastFinishedPulling="2025-10-14 08:17:13.482572794 +0000 UTC m=+5250.066619421" observedRunningTime="2025-10-14 08:17:13.784809589 +0000 UTC m=+5250.368856256" watchObservedRunningTime="2025-10-14 08:17:13.801947084 +0000 UTC m=+5250.385993721" Oct 14 08:17:14 crc kubenswrapper[5018]: I1014 08:17:14.776742 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerStarted","Data":"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10"} Oct 14 08:17:15 crc kubenswrapper[5018]: I1014 08:17:15.791032 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerID="c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10" exitCode=0 Oct 14 08:17:15 crc kubenswrapper[5018]: I1014 08:17:15.791264 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerDied","Data":"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10"} Oct 14 08:17:16 crc kubenswrapper[5018]: I1014 08:17:16.806162 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerStarted","Data":"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb"} Oct 14 08:17:16 crc kubenswrapper[5018]: I1014 08:17:16.826314 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-th7t9" podStartSLOduration=2.110565889 podStartE2EDuration="4.826293465s" podCreationTimestamp="2025-10-14 08:17:12 +0000 UTC" firstStartedPulling="2025-10-14 08:17:13.765603794 +0000 UTC m=+5250.349650421" lastFinishedPulling="2025-10-14 08:17:16.48133136 +0000 UTC m=+5253.065377997" observedRunningTime="2025-10-14 08:17:16.820995195 +0000 UTC m=+5253.405041842" watchObservedRunningTime="2025-10-14 08:17:16.826293465 +0000 UTC m=+5253.410340092" Oct 14 08:17:18 crc kubenswrapper[5018]: I1014 08:17:18.605351 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:17:18 crc kubenswrapper[5018]: E1014 08:17:18.605579 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:17:19 crc kubenswrapper[5018]: I1014 08:17:19.994647 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:19 crc kubenswrapper[5018]: I1014 08:17:19.995244 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:20 crc kubenswrapper[5018]: I1014 08:17:20.075808 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:20 crc kubenswrapper[5018]: I1014 08:17:20.905028 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:21 crc kubenswrapper[5018]: I1014 08:17:21.219436 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:22 crc kubenswrapper[5018]: I1014 08:17:22.800107 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:22 crc kubenswrapper[5018]: I1014 08:17:22.801368 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:22 crc kubenswrapper[5018]: I1014 08:17:22.859918 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cd6jd" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="registry-server" containerID="cri-o://6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9" gracePeriod=2 Oct 14 08:17:22 crc kubenswrapper[5018]: I1014 08:17:22.867151 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.405216 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.464437 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities\") pod \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.464622 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content\") pod \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.464687 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm8hb\" (UniqueName: \"kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb\") pod \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\" (UID: \"736bee71-4d68-4f1e-8339-e8eb9dbc2384\") " Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.465360 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities" (OuterVolumeSpecName: "utilities") pod "736bee71-4d68-4f1e-8339-e8eb9dbc2384" (UID: "736bee71-4d68-4f1e-8339-e8eb9dbc2384"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.472217 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb" (OuterVolumeSpecName: "kube-api-access-fm8hb") pod "736bee71-4d68-4f1e-8339-e8eb9dbc2384" (UID: "736bee71-4d68-4f1e-8339-e8eb9dbc2384"). InnerVolumeSpecName "kube-api-access-fm8hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.511317 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "736bee71-4d68-4f1e-8339-e8eb9dbc2384" (UID: "736bee71-4d68-4f1e-8339-e8eb9dbc2384"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.565926 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.565958 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm8hb\" (UniqueName: \"kubernetes.io/projected/736bee71-4d68-4f1e-8339-e8eb9dbc2384-kube-api-access-fm8hb\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.565969 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736bee71-4d68-4f1e-8339-e8eb9dbc2384-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.879347 5018 generic.go:334] "Generic (PLEG): container finished" podID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerID="6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9" exitCode=0 Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.879408 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd6jd" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.879577 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerDied","Data":"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9"} Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.880052 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd6jd" event={"ID":"736bee71-4d68-4f1e-8339-e8eb9dbc2384","Type":"ContainerDied","Data":"2874104d183fd3c22426266a7563af96adaaae287d594905d8e6b0ca48e7a831"} Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.880128 5018 scope.go:117] "RemoveContainer" containerID="6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.900840 5018 scope.go:117] "RemoveContainer" containerID="9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.934015 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.934693 5018 scope.go:117] "RemoveContainer" containerID="81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.941170 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cd6jd"] Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.962917 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.969214 5018 scope.go:117] "RemoveContainer" containerID="6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9" Oct 14 08:17:23 crc kubenswrapper[5018]: E1014 08:17:23.970852 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9\": container with ID starting with 6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9 not found: ID does not exist" containerID="6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.970902 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9"} err="failed to get container status \"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9\": rpc error: code = NotFound desc = could not find container \"6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9\": container with ID starting with 6d89494f4b2b2e6a3dacb7e4017dc8f82fa06f5aaa2feeaef9baf318f106fec9 not found: ID does not exist" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.970931 5018 scope.go:117] "RemoveContainer" containerID="9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b" Oct 14 08:17:23 crc kubenswrapper[5018]: E1014 08:17:23.971521 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b\": container with ID starting with 9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b not found: ID does not exist" containerID="9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.971579 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b"} err="failed to get container status \"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b\": rpc error: code = NotFound desc = could not find container \"9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b\": container with ID starting with 9980c0de538075ab28bc50fd5e5c71d18d7e278201b094767582dd108538553b not found: ID does not exist" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.971633 5018 scope.go:117] "RemoveContainer" containerID="81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5" Oct 14 08:17:23 crc kubenswrapper[5018]: E1014 08:17:23.972057 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5\": container with ID starting with 81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5 not found: ID does not exist" containerID="81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5" Oct 14 08:17:23 crc kubenswrapper[5018]: I1014 08:17:23.972108 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5"} err="failed to get container status \"81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5\": rpc error: code = NotFound desc = could not find container \"81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5\": container with ID starting with 81f0ec39b163c3296de942de25051f0cb9b9222847941a8af60dfcb0138cd1e5 not found: ID does not exist" Oct 14 08:17:24 crc kubenswrapper[5018]: I1014 08:17:24.623211 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" path="/var/lib/kubelet/pods/736bee71-4d68-4f1e-8339-e8eb9dbc2384/volumes" Oct 14 08:17:26 crc kubenswrapper[5018]: I1014 08:17:26.821521 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:26 crc kubenswrapper[5018]: I1014 08:17:26.822269 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-th7t9" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="registry-server" containerID="cri-o://5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb" gracePeriod=2 Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.275921 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.426335 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities\") pod \"ce007799-de66-46b7-a2f1-1f250a5ea652\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.426453 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content\") pod \"ce007799-de66-46b7-a2f1-1f250a5ea652\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.426530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnbbf\" (UniqueName: \"kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf\") pod \"ce007799-de66-46b7-a2f1-1f250a5ea652\" (UID: \"ce007799-de66-46b7-a2f1-1f250a5ea652\") " Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.428042 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities" (OuterVolumeSpecName: "utilities") pod "ce007799-de66-46b7-a2f1-1f250a5ea652" (UID: "ce007799-de66-46b7-a2f1-1f250a5ea652"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.435001 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf" (OuterVolumeSpecName: "kube-api-access-rnbbf") pod "ce007799-de66-46b7-a2f1-1f250a5ea652" (UID: "ce007799-de66-46b7-a2f1-1f250a5ea652"). InnerVolumeSpecName "kube-api-access-rnbbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.531577 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnbbf\" (UniqueName: \"kubernetes.io/projected/ce007799-de66-46b7-a2f1-1f250a5ea652-kube-api-access-rnbbf\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.531623 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.571574 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce007799-de66-46b7-a2f1-1f250a5ea652" (UID: "ce007799-de66-46b7-a2f1-1f250a5ea652"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.633242 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce007799-de66-46b7-a2f1-1f250a5ea652-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.921755 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerID="5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb" exitCode=0 Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.921838 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerDied","Data":"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb"} Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.921893 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-th7t9" event={"ID":"ce007799-de66-46b7-a2f1-1f250a5ea652","Type":"ContainerDied","Data":"086110c0400db100c9537c324f4979a5c9fd7a53a39abfacdbd55ca2a644937e"} Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.921910 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-th7t9" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.921932 5018 scope.go:117] "RemoveContainer" containerID="5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.964770 5018 scope.go:117] "RemoveContainer" containerID="c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10" Oct 14 08:17:27 crc kubenswrapper[5018]: I1014 08:17:27.999695 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.005820 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-th7t9"] Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.007798 5018 scope.go:117] "RemoveContainer" containerID="5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.043757 5018 scope.go:117] "RemoveContainer" containerID="5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb" Oct 14 08:17:28 crc kubenswrapper[5018]: E1014 08:17:28.045078 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb\": container with ID starting with 5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb not found: ID does not exist" containerID="5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.045107 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb"} err="failed to get container status \"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb\": rpc error: code = NotFound desc = could not find container \"5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb\": container with ID starting with 5ee0299af63888f75411cbc3bff59ac47f35bab4d7792a3eee2ba34ab26d7cdb not found: ID does not exist" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.045132 5018 scope.go:117] "RemoveContainer" containerID="c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10" Oct 14 08:17:28 crc kubenswrapper[5018]: E1014 08:17:28.045528 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10\": container with ID starting with c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10 not found: ID does not exist" containerID="c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.045551 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10"} err="failed to get container status \"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10\": rpc error: code = NotFound desc = could not find container \"c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10\": container with ID starting with c31043a830d2e8eec4b01e530ec76d90d15693e9b2af599449213ce1846eed10 not found: ID does not exist" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.045574 5018 scope.go:117] "RemoveContainer" containerID="5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9" Oct 14 08:17:28 crc kubenswrapper[5018]: E1014 08:17:28.045978 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9\": container with ID starting with 5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9 not found: ID does not exist" containerID="5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.045997 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9"} err="failed to get container status \"5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9\": rpc error: code = NotFound desc = could not find container \"5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9\": container with ID starting with 5e60a38725f4d1258546d5ddc11749c4503670563b2070af320f5edca668b2c9 not found: ID does not exist" Oct 14 08:17:28 crc kubenswrapper[5018]: I1014 08:17:28.617129 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" path="/var/lib/kubelet/pods/ce007799-de66-46b7-a2f1-1f250a5ea652/volumes" Oct 14 08:17:30 crc kubenswrapper[5018]: I1014 08:17:30.604946 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:17:30 crc kubenswrapper[5018]: E1014 08:17:30.605547 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:17:45 crc kubenswrapper[5018]: I1014 08:17:45.606377 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:17:45 crc kubenswrapper[5018]: E1014 08:17:45.607512 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:18:00 crc kubenswrapper[5018]: I1014 08:18:00.606044 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:18:00 crc kubenswrapper[5018]: E1014 08:18:00.607285 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:18:11 crc kubenswrapper[5018]: I1014 08:18:11.605729 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:18:12 crc kubenswrapper[5018]: I1014 08:18:12.334211 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298"} Oct 14 08:20:32 crc kubenswrapper[5018]: I1014 08:20:32.464016 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:20:32 crc kubenswrapper[5018]: I1014 08:20:32.464830 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:21:02 crc kubenswrapper[5018]: I1014 08:21:02.463936 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:21:02 crc kubenswrapper[5018]: I1014 08:21:02.464679 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:21:32 crc kubenswrapper[5018]: I1014 08:21:32.463439 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:21:32 crc kubenswrapper[5018]: I1014 08:21:32.464305 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:21:32 crc kubenswrapper[5018]: I1014 08:21:32.464402 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:21:32 crc kubenswrapper[5018]: I1014 08:21:32.465609 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:21:32 crc kubenswrapper[5018]: I1014 08:21:32.465771 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298" gracePeriod=600 Oct 14 08:21:33 crc kubenswrapper[5018]: I1014 08:21:33.326640 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298" exitCode=0 Oct 14 08:21:33 crc kubenswrapper[5018]: I1014 08:21:33.326660 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298"} Oct 14 08:21:33 crc kubenswrapper[5018]: I1014 08:21:33.327027 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78"} Oct 14 08:21:33 crc kubenswrapper[5018]: I1014 08:21:33.327067 5018 scope.go:117] "RemoveContainer" containerID="1532aa60343212bbb3e6ffe78eb8f5e38a32887d48f15f2de2314cb98ebc020f" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.744112 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745278 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="extract-utilities" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745308 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="extract-utilities" Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745347 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="extract-content" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745364 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="extract-content" Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745420 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745442 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745482 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="extract-utilities" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745496 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="extract-utilities" Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745511 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="extract-content" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745524 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="extract-content" Oct 14 08:23:23 crc kubenswrapper[5018]: E1014 08:23:23.745545 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745558 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745884 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce007799-de66-46b7-a2f1-1f250a5ea652" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.745962 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="736bee71-4d68-4f1e-8339-e8eb9dbc2384" containerName="registry-server" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.748669 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.762164 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.866871 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.866926 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlfkx\" (UniqueName: \"kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.867123 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.969202 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlfkx\" (UniqueName: \"kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.969297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.969429 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.969749 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.969906 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:23 crc kubenswrapper[5018]: I1014 08:23:23.988760 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlfkx\" (UniqueName: \"kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx\") pod \"community-operators-7zxdl\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:24 crc kubenswrapper[5018]: I1014 08:23:24.128796 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:24 crc kubenswrapper[5018]: I1014 08:23:24.625841 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:25 crc kubenswrapper[5018]: I1014 08:23:25.472069 5018 generic.go:334] "Generic (PLEG): container finished" podID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerID="f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b" exitCode=0 Oct 14 08:23:25 crc kubenswrapper[5018]: I1014 08:23:25.472119 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerDied","Data":"f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b"} Oct 14 08:23:25 crc kubenswrapper[5018]: I1014 08:23:25.472154 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerStarted","Data":"ac06d04418865d363a1b90b6138263d7fe26412f700415c7d72f7b388480824a"} Oct 14 08:23:25 crc kubenswrapper[5018]: I1014 08:23:25.475793 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:23:27 crc kubenswrapper[5018]: I1014 08:23:27.516395 5018 generic.go:334] "Generic (PLEG): container finished" podID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerID="10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258" exitCode=0 Oct 14 08:23:27 crc kubenswrapper[5018]: I1014 08:23:27.516899 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerDied","Data":"10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258"} Oct 14 08:23:28 crc kubenswrapper[5018]: I1014 08:23:28.528323 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerStarted","Data":"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41"} Oct 14 08:23:28 crc kubenswrapper[5018]: I1014 08:23:28.548902 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7zxdl" podStartSLOduration=2.8279767749999998 podStartE2EDuration="5.548874506s" podCreationTimestamp="2025-10-14 08:23:23 +0000 UTC" firstStartedPulling="2025-10-14 08:23:25.475422176 +0000 UTC m=+5622.059468813" lastFinishedPulling="2025-10-14 08:23:28.196319887 +0000 UTC m=+5624.780366544" observedRunningTime="2025-10-14 08:23:28.542972789 +0000 UTC m=+5625.127019426" watchObservedRunningTime="2025-10-14 08:23:28.548874506 +0000 UTC m=+5625.132921163" Oct 14 08:23:32 crc kubenswrapper[5018]: I1014 08:23:32.464065 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:23:32 crc kubenswrapper[5018]: I1014 08:23:32.464509 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:23:34 crc kubenswrapper[5018]: I1014 08:23:34.129713 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:34 crc kubenswrapper[5018]: I1014 08:23:34.130248 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:34 crc kubenswrapper[5018]: I1014 08:23:34.192731 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:34 crc kubenswrapper[5018]: I1014 08:23:34.658927 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:34 crc kubenswrapper[5018]: I1014 08:23:34.731753 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:36 crc kubenswrapper[5018]: I1014 08:23:36.607021 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7zxdl" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="registry-server" containerID="cri-o://65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41" gracePeriod=2 Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.047537 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.129880 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content\") pod \"f9eedc5f-9cab-416e-b337-7e321a8ea254\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.129948 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlfkx\" (UniqueName: \"kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx\") pod \"f9eedc5f-9cab-416e-b337-7e321a8ea254\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.130085 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities\") pod \"f9eedc5f-9cab-416e-b337-7e321a8ea254\" (UID: \"f9eedc5f-9cab-416e-b337-7e321a8ea254\") " Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.132022 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities" (OuterVolumeSpecName: "utilities") pod "f9eedc5f-9cab-416e-b337-7e321a8ea254" (UID: "f9eedc5f-9cab-416e-b337-7e321a8ea254"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.136564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx" (OuterVolumeSpecName: "kube-api-access-rlfkx") pod "f9eedc5f-9cab-416e-b337-7e321a8ea254" (UID: "f9eedc5f-9cab-416e-b337-7e321a8ea254"). InnerVolumeSpecName "kube-api-access-rlfkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.199169 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9eedc5f-9cab-416e-b337-7e321a8ea254" (UID: "f9eedc5f-9cab-416e-b337-7e321a8ea254"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.231069 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.231103 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlfkx\" (UniqueName: \"kubernetes.io/projected/f9eedc5f-9cab-416e-b337-7e321a8ea254-kube-api-access-rlfkx\") on node \"crc\" DevicePath \"\"" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.231113 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9eedc5f-9cab-416e-b337-7e321a8ea254-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.633457 5018 generic.go:334] "Generic (PLEG): container finished" podID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerID="65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41" exitCode=0 Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.633529 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerDied","Data":"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41"} Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.633591 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zxdl" event={"ID":"f9eedc5f-9cab-416e-b337-7e321a8ea254","Type":"ContainerDied","Data":"ac06d04418865d363a1b90b6138263d7fe26412f700415c7d72f7b388480824a"} Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.633648 5018 scope.go:117] "RemoveContainer" containerID="65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.633608 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zxdl" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.663261 5018 scope.go:117] "RemoveContainer" containerID="10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.694921 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.703847 5018 scope.go:117] "RemoveContainer" containerID="f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.707764 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7zxdl"] Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.727820 5018 scope.go:117] "RemoveContainer" containerID="65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41" Oct 14 08:23:37 crc kubenswrapper[5018]: E1014 08:23:37.728107 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41\": container with ID starting with 65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41 not found: ID does not exist" containerID="65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.728136 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41"} err="failed to get container status \"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41\": rpc error: code = NotFound desc = could not find container \"65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41\": container with ID starting with 65bc8dcb26f617b552d978b3e790e2efde9cf5a8f5edc428e967229c3f732c41 not found: ID does not exist" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.728158 5018 scope.go:117] "RemoveContainer" containerID="10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258" Oct 14 08:23:37 crc kubenswrapper[5018]: E1014 08:23:37.728530 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258\": container with ID starting with 10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258 not found: ID does not exist" containerID="10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.728735 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258"} err="failed to get container status \"10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258\": rpc error: code = NotFound desc = could not find container \"10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258\": container with ID starting with 10aaa48e5f47f38db018417c192f9250bfc8232b575ea2d09de1f59573da8258 not found: ID does not exist" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.728867 5018 scope.go:117] "RemoveContainer" containerID="f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b" Oct 14 08:23:37 crc kubenswrapper[5018]: E1014 08:23:37.729254 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b\": container with ID starting with f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b not found: ID does not exist" containerID="f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b" Oct 14 08:23:37 crc kubenswrapper[5018]: I1014 08:23:37.729298 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b"} err="failed to get container status \"f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b\": rpc error: code = NotFound desc = could not find container \"f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b\": container with ID starting with f9fe11ebddbb01f401d5a537f924df82145a98342d2eae6446e4dfdc6bf5b75b not found: ID does not exist" Oct 14 08:23:38 crc kubenswrapper[5018]: I1014 08:23:38.620112 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" path="/var/lib/kubelet/pods/f9eedc5f-9cab-416e-b337-7e321a8ea254/volumes" Oct 14 08:24:02 crc kubenswrapper[5018]: I1014 08:24:02.463985 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:24:02 crc kubenswrapper[5018]: I1014 08:24:02.464615 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:24:32 crc kubenswrapper[5018]: I1014 08:24:32.463822 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:24:32 crc kubenswrapper[5018]: I1014 08:24:32.464731 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:24:32 crc kubenswrapper[5018]: I1014 08:24:32.464809 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:24:32 crc kubenswrapper[5018]: I1014 08:24:32.465874 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:24:32 crc kubenswrapper[5018]: I1014 08:24:32.465978 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" gracePeriod=600 Oct 14 08:24:32 crc kubenswrapper[5018]: E1014 08:24:32.599033 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:24:33 crc kubenswrapper[5018]: I1014 08:24:33.187477 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" exitCode=0 Oct 14 08:24:33 crc kubenswrapper[5018]: I1014 08:24:33.187572 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78"} Oct 14 08:24:33 crc kubenswrapper[5018]: I1014 08:24:33.187729 5018 scope.go:117] "RemoveContainer" containerID="bc7e0c1692314ef02b7895bd10635a017d4625849bd9676aa5dee717b431b298" Oct 14 08:24:33 crc kubenswrapper[5018]: I1014 08:24:33.188685 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:24:33 crc kubenswrapper[5018]: E1014 08:24:33.189309 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:24:46 crc kubenswrapper[5018]: I1014 08:24:46.604796 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:24:46 crc kubenswrapper[5018]: E1014 08:24:46.605540 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:24:58 crc kubenswrapper[5018]: I1014 08:24:58.606782 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:24:58 crc kubenswrapper[5018]: E1014 08:24:58.608340 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:25:12 crc kubenswrapper[5018]: I1014 08:25:12.605599 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:25:12 crc kubenswrapper[5018]: E1014 08:25:12.606761 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:25:23 crc kubenswrapper[5018]: I1014 08:25:23.605782 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:25:23 crc kubenswrapper[5018]: E1014 08:25:23.607081 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:25:38 crc kubenswrapper[5018]: I1014 08:25:38.606289 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:25:38 crc kubenswrapper[5018]: E1014 08:25:38.607676 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:25:51 crc kubenswrapper[5018]: I1014 08:25:51.605414 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:25:51 crc kubenswrapper[5018]: E1014 08:25:51.607235 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.265752 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:00 crc kubenswrapper[5018]: E1014 08:26:00.267129 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="registry-server" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.267156 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="registry-server" Oct 14 08:26:00 crc kubenswrapper[5018]: E1014 08:26:00.267181 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="extract-utilities" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.267193 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="extract-utilities" Oct 14 08:26:00 crc kubenswrapper[5018]: E1014 08:26:00.267218 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="extract-content" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.267231 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="extract-content" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.267519 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9eedc5f-9cab-416e-b337-7e321a8ea254" containerName="registry-server" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.269832 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.344739 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.423329 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.423678 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw82l\" (UniqueName: \"kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.423853 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.525774 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.526327 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.526613 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw82l\" (UniqueName: \"kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.526907 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.526426 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.556402 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw82l\" (UniqueName: \"kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l\") pod \"redhat-marketplace-fffwk\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.611031 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:00 crc kubenswrapper[5018]: I1014 08:26:00.877821 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:01 crc kubenswrapper[5018]: I1014 08:26:01.044507 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerStarted","Data":"0568d4e613e54b4afbd301ee25d1b67bd16bc557aa0b0f1066c430e9d5ec8a5a"} Oct 14 08:26:02 crc kubenswrapper[5018]: I1014 08:26:02.086148 5018 generic.go:334] "Generic (PLEG): container finished" podID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerID="6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514" exitCode=0 Oct 14 08:26:02 crc kubenswrapper[5018]: I1014 08:26:02.086219 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerDied","Data":"6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514"} Oct 14 08:26:02 crc kubenswrapper[5018]: I1014 08:26:02.606592 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:26:02 crc kubenswrapper[5018]: E1014 08:26:02.607567 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:26:04 crc kubenswrapper[5018]: I1014 08:26:04.107364 5018 generic.go:334] "Generic (PLEG): container finished" podID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerID="e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9" exitCode=0 Oct 14 08:26:04 crc kubenswrapper[5018]: I1014 08:26:04.107449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerDied","Data":"e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9"} Oct 14 08:26:05 crc kubenswrapper[5018]: I1014 08:26:05.125021 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerStarted","Data":"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8"} Oct 14 08:26:05 crc kubenswrapper[5018]: I1014 08:26:05.162415 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fffwk" podStartSLOduration=2.756151638 podStartE2EDuration="5.162387604s" podCreationTimestamp="2025-10-14 08:26:00 +0000 UTC" firstStartedPulling="2025-10-14 08:26:02.091519506 +0000 UTC m=+5778.675566183" lastFinishedPulling="2025-10-14 08:26:04.497755492 +0000 UTC m=+5781.081802149" observedRunningTime="2025-10-14 08:26:05.150658741 +0000 UTC m=+5781.734705418" watchObservedRunningTime="2025-10-14 08:26:05.162387604 +0000 UTC m=+5781.746434271" Oct 14 08:26:10 crc kubenswrapper[5018]: I1014 08:26:10.624068 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:10 crc kubenswrapper[5018]: I1014 08:26:10.624674 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:10 crc kubenswrapper[5018]: I1014 08:26:10.698871 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:11 crc kubenswrapper[5018]: I1014 08:26:11.249224 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:11 crc kubenswrapper[5018]: I1014 08:26:11.313561 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.209028 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fffwk" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="registry-server" containerID="cri-o://4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8" gracePeriod=2 Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.712651 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.769825 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities\") pod \"6f2aba43-3da3-444b-a7bb-5db77c136348\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.769999 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw82l\" (UniqueName: \"kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l\") pod \"6f2aba43-3da3-444b-a7bb-5db77c136348\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.770057 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content\") pod \"6f2aba43-3da3-444b-a7bb-5db77c136348\" (UID: \"6f2aba43-3da3-444b-a7bb-5db77c136348\") " Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.772508 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities" (OuterVolumeSpecName: "utilities") pod "6f2aba43-3da3-444b-a7bb-5db77c136348" (UID: "6f2aba43-3da3-444b-a7bb-5db77c136348"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.780084 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l" (OuterVolumeSpecName: "kube-api-access-gw82l") pod "6f2aba43-3da3-444b-a7bb-5db77c136348" (UID: "6f2aba43-3da3-444b-a7bb-5db77c136348"). InnerVolumeSpecName "kube-api-access-gw82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.790472 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f2aba43-3da3-444b-a7bb-5db77c136348" (UID: "6f2aba43-3da3-444b-a7bb-5db77c136348"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.871502 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.871540 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f2aba43-3da3-444b-a7bb-5db77c136348-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:13 crc kubenswrapper[5018]: I1014 08:26:13.871554 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw82l\" (UniqueName: \"kubernetes.io/projected/6f2aba43-3da3-444b-a7bb-5db77c136348-kube-api-access-gw82l\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.220761 5018 generic.go:334] "Generic (PLEG): container finished" podID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerID="4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8" exitCode=0 Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.220846 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fffwk" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.220864 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerDied","Data":"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8"} Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.220950 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fffwk" event={"ID":"6f2aba43-3da3-444b-a7bb-5db77c136348","Type":"ContainerDied","Data":"0568d4e613e54b4afbd301ee25d1b67bd16bc557aa0b0f1066c430e9d5ec8a5a"} Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.220983 5018 scope.go:117] "RemoveContainer" containerID="4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.259775 5018 scope.go:117] "RemoveContainer" containerID="e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.268093 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.275180 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fffwk"] Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.287822 5018 scope.go:117] "RemoveContainer" containerID="6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.338458 5018 scope.go:117] "RemoveContainer" containerID="4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8" Oct 14 08:26:14 crc kubenswrapper[5018]: E1014 08:26:14.338965 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8\": container with ID starting with 4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8 not found: ID does not exist" containerID="4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.339014 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8"} err="failed to get container status \"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8\": rpc error: code = NotFound desc = could not find container \"4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8\": container with ID starting with 4e43b3c856d200cfc04314ba56c9e659232cf48e4e5e5431b18e018c585e28c8 not found: ID does not exist" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.339047 5018 scope.go:117] "RemoveContainer" containerID="e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9" Oct 14 08:26:14 crc kubenswrapper[5018]: E1014 08:26:14.339754 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9\": container with ID starting with e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9 not found: ID does not exist" containerID="e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.339802 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9"} err="failed to get container status \"e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9\": rpc error: code = NotFound desc = could not find container \"e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9\": container with ID starting with e3ec5f49a42be7e90a4970f18a24b1e434fbe052d04ea8cd81221f6f6feabee9 not found: ID does not exist" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.339832 5018 scope.go:117] "RemoveContainer" containerID="6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514" Oct 14 08:26:14 crc kubenswrapper[5018]: E1014 08:26:14.340597 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514\": container with ID starting with 6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514 not found: ID does not exist" containerID="6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.340654 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514"} err="failed to get container status \"6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514\": rpc error: code = NotFound desc = could not find container \"6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514\": container with ID starting with 6fd5c5545703d27c2142b094a071a1f5f7e761d688d3e1936dcfc805657f9514 not found: ID does not exist" Oct 14 08:26:14 crc kubenswrapper[5018]: I1014 08:26:14.618759 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" path="/var/lib/kubelet/pods/6f2aba43-3da3-444b-a7bb-5db77c136348/volumes" Oct 14 08:26:16 crc kubenswrapper[5018]: I1014 08:26:16.605835 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:26:16 crc kubenswrapper[5018]: E1014 08:26:16.606721 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:26:27 crc kubenswrapper[5018]: I1014 08:26:27.604895 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:26:27 crc kubenswrapper[5018]: E1014 08:26:27.606094 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:26:38 crc kubenswrapper[5018]: I1014 08:26:38.605399 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:26:38 crc kubenswrapper[5018]: E1014 08:26:38.606383 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:26:49 crc kubenswrapper[5018]: I1014 08:26:49.605429 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:26:49 crc kubenswrapper[5018]: E1014 08:26:49.606504 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:27:00 crc kubenswrapper[5018]: I1014 08:27:00.605919 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:27:00 crc kubenswrapper[5018]: E1014 08:27:00.606694 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:27:13 crc kubenswrapper[5018]: I1014 08:27:13.605546 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:27:13 crc kubenswrapper[5018]: E1014 08:27:13.606677 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.099525 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:21 crc kubenswrapper[5018]: E1014 08:27:21.100435 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="registry-server" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.100452 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="registry-server" Oct 14 08:27:21 crc kubenswrapper[5018]: E1014 08:27:21.100465 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="extract-content" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.100474 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="extract-content" Oct 14 08:27:21 crc kubenswrapper[5018]: E1014 08:27:21.100499 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="extract-utilities" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.100508 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="extract-utilities" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.100754 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f2aba43-3da3-444b-a7bb-5db77c136348" containerName="registry-server" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.102030 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.123195 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.299909 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vhcc\" (UniqueName: \"kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.300018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.300055 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.401917 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.402009 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.402042 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vhcc\" (UniqueName: \"kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.402884 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.402946 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.441128 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vhcc\" (UniqueName: \"kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc\") pod \"redhat-operators-dtcg7\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.459771 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.716638 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:21 crc kubenswrapper[5018]: I1014 08:27:21.849834 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerStarted","Data":"85b53e2d9e3efae1668fbc009de15786ba614b8c112824c6e6c7a2672576d63a"} Oct 14 08:27:22 crc kubenswrapper[5018]: I1014 08:27:22.860158 5018 generic.go:334] "Generic (PLEG): container finished" podID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerID="1542282a89a23a44f384197f8c3fc8e2f4bdacbb48b918bf86765243f9e25d22" exitCode=0 Oct 14 08:27:22 crc kubenswrapper[5018]: I1014 08:27:22.860242 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerDied","Data":"1542282a89a23a44f384197f8c3fc8e2f4bdacbb48b918bf86765243f9e25d22"} Oct 14 08:27:23 crc kubenswrapper[5018]: I1014 08:27:23.871694 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerStarted","Data":"c9585659bb37b2198dfb207da2e61c41c96bc4dc83b470d3fc11c35c1c5cdacd"} Oct 14 08:27:24 crc kubenswrapper[5018]: I1014 08:27:24.884917 5018 generic.go:334] "Generic (PLEG): container finished" podID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerID="c9585659bb37b2198dfb207da2e61c41c96bc4dc83b470d3fc11c35c1c5cdacd" exitCode=0 Oct 14 08:27:24 crc kubenswrapper[5018]: I1014 08:27:24.885055 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerDied","Data":"c9585659bb37b2198dfb207da2e61c41c96bc4dc83b470d3fc11c35c1c5cdacd"} Oct 14 08:27:25 crc kubenswrapper[5018]: I1014 08:27:25.606084 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:27:25 crc kubenswrapper[5018]: E1014 08:27:25.606498 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:27:26 crc kubenswrapper[5018]: I1014 08:27:26.905278 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerStarted","Data":"5cd523cba377fd487712d07c88c9965638070d41087d4cfea5ce502a3248d2d9"} Oct 14 08:27:26 crc kubenswrapper[5018]: I1014 08:27:26.945273 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dtcg7" podStartSLOduration=2.522602297 podStartE2EDuration="5.945245471s" podCreationTimestamp="2025-10-14 08:27:21 +0000 UTC" firstStartedPulling="2025-10-14 08:27:22.864726317 +0000 UTC m=+5859.448772984" lastFinishedPulling="2025-10-14 08:27:26.287369521 +0000 UTC m=+5862.871416158" observedRunningTime="2025-10-14 08:27:26.935213077 +0000 UTC m=+5863.519259744" watchObservedRunningTime="2025-10-14 08:27:26.945245471 +0000 UTC m=+5863.529292138" Oct 14 08:27:31 crc kubenswrapper[5018]: I1014 08:27:31.460296 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:31 crc kubenswrapper[5018]: I1014 08:27:31.460710 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:32 crc kubenswrapper[5018]: I1014 08:27:32.537306 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtcg7" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="registry-server" probeResult="failure" output=< Oct 14 08:27:32 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 08:27:32 crc kubenswrapper[5018]: > Oct 14 08:27:37 crc kubenswrapper[5018]: I1014 08:27:37.604721 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:27:37 crc kubenswrapper[5018]: E1014 08:27:37.605870 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:27:41 crc kubenswrapper[5018]: I1014 08:27:41.544400 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:41 crc kubenswrapper[5018]: I1014 08:27:41.615390 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:41 crc kubenswrapper[5018]: I1014 08:27:41.794081 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:43 crc kubenswrapper[5018]: I1014 08:27:43.063036 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dtcg7" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="registry-server" containerID="cri-o://5cd523cba377fd487712d07c88c9965638070d41087d4cfea5ce502a3248d2d9" gracePeriod=2 Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.086004 5018 generic.go:334] "Generic (PLEG): container finished" podID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerID="5cd523cba377fd487712d07c88c9965638070d41087d4cfea5ce502a3248d2d9" exitCode=0 Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.086057 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerDied","Data":"5cd523cba377fd487712d07c88c9965638070d41087d4cfea5ce502a3248d2d9"} Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.086091 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtcg7" event={"ID":"0171f3a9-e52c-43f8-ba5d-f89fb00cf287","Type":"ContainerDied","Data":"85b53e2d9e3efae1668fbc009de15786ba614b8c112824c6e6c7a2672576d63a"} Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.086107 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85b53e2d9e3efae1668fbc009de15786ba614b8c112824c6e6c7a2672576d63a" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.091380 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.143168 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities\") pod \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.143870 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content\") pod \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.143961 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vhcc\" (UniqueName: \"kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc\") pod \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\" (UID: \"0171f3a9-e52c-43f8-ba5d-f89fb00cf287\") " Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.144392 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities" (OuterVolumeSpecName: "utilities") pod "0171f3a9-e52c-43f8-ba5d-f89fb00cf287" (UID: "0171f3a9-e52c-43f8-ba5d-f89fb00cf287"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.144891 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.154278 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc" (OuterVolumeSpecName: "kube-api-access-8vhcc") pod "0171f3a9-e52c-43f8-ba5d-f89fb00cf287" (UID: "0171f3a9-e52c-43f8-ba5d-f89fb00cf287"). InnerVolumeSpecName "kube-api-access-8vhcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.246976 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vhcc\" (UniqueName: \"kubernetes.io/projected/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-kube-api-access-8vhcc\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.266721 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0171f3a9-e52c-43f8-ba5d-f89fb00cf287" (UID: "0171f3a9-e52c-43f8-ba5d-f89fb00cf287"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:27:44 crc kubenswrapper[5018]: I1014 08:27:44.348945 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0171f3a9-e52c-43f8-ba5d-f89fb00cf287-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:45 crc kubenswrapper[5018]: I1014 08:27:45.101841 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtcg7" Oct 14 08:27:45 crc kubenswrapper[5018]: I1014 08:27:45.142448 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:45 crc kubenswrapper[5018]: I1014 08:27:45.150414 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dtcg7"] Oct 14 08:27:46 crc kubenswrapper[5018]: I1014 08:27:46.621725 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" path="/var/lib/kubelet/pods/0171f3a9-e52c-43f8-ba5d-f89fb00cf287/volumes" Oct 14 08:27:52 crc kubenswrapper[5018]: I1014 08:27:52.605551 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:27:52 crc kubenswrapper[5018]: E1014 08:27:52.606306 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.508074 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-dtwvt"] Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.514641 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-dtwvt"] Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.622039 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0ac94c-3256-4e42-9520-a7af1cc2266e" path="/var/lib/kubelet/pods/9f0ac94c-3256-4e42-9520-a7af1cc2266e/volumes" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.623457 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-2mj5x"] Oct 14 08:28:00 crc kubenswrapper[5018]: E1014 08:28:00.624166 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="registry-server" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.624199 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="registry-server" Oct 14 08:28:00 crc kubenswrapper[5018]: E1014 08:28:00.624238 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="extract-content" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.624252 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="extract-content" Oct 14 08:28:00 crc kubenswrapper[5018]: E1014 08:28:00.624282 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="extract-utilities" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.624293 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="extract-utilities" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.624590 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0171f3a9-e52c-43f8-ba5d-f89fb00cf287" containerName="registry-server" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.625830 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.628162 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.628217 5018 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5sfhc" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.628284 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.628162 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.636431 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2mj5x"] Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.744815 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.744866 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.744922 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb8gq\" (UniqueName: \"kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.846581 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.846674 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.846722 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb8gq\" (UniqueName: \"kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.847174 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.847648 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.875718 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb8gq\" (UniqueName: \"kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq\") pod \"crc-storage-crc-2mj5x\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:00 crc kubenswrapper[5018]: I1014 08:28:00.960701 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:01 crc kubenswrapper[5018]: I1014 08:28:01.439022 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2mj5x"] Oct 14 08:28:02 crc kubenswrapper[5018]: I1014 08:28:02.253083 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2mj5x" event={"ID":"d635d90c-f19b-4646-bde8-bffe1a0da42a","Type":"ContainerStarted","Data":"b350ff55fd93c69942fd77ca393744433abc13f145b494dd68a72c35f873eee4"} Oct 14 08:28:02 crc kubenswrapper[5018]: I1014 08:28:02.253439 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2mj5x" event={"ID":"d635d90c-f19b-4646-bde8-bffe1a0da42a","Type":"ContainerStarted","Data":"7bc1ef92d49c205283824aae80735deb651b4cdbf7ed75f690948d303e83e8bb"} Oct 14 08:28:02 crc kubenswrapper[5018]: I1014 08:28:02.272772 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-2mj5x" podStartSLOduration=1.751662063 podStartE2EDuration="2.272751228s" podCreationTimestamp="2025-10-14 08:28:00 +0000 UTC" firstStartedPulling="2025-10-14 08:28:01.453733173 +0000 UTC m=+5898.037779840" lastFinishedPulling="2025-10-14 08:28:01.974822368 +0000 UTC m=+5898.558869005" observedRunningTime="2025-10-14 08:28:02.267066527 +0000 UTC m=+5898.851113154" watchObservedRunningTime="2025-10-14 08:28:02.272751228 +0000 UTC m=+5898.856797865" Oct 14 08:28:03 crc kubenswrapper[5018]: I1014 08:28:03.266428 5018 generic.go:334] "Generic (PLEG): container finished" podID="d635d90c-f19b-4646-bde8-bffe1a0da42a" containerID="b350ff55fd93c69942fd77ca393744433abc13f145b494dd68a72c35f873eee4" exitCode=0 Oct 14 08:28:03 crc kubenswrapper[5018]: I1014 08:28:03.266532 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2mj5x" event={"ID":"d635d90c-f19b-4646-bde8-bffe1a0da42a","Type":"ContainerDied","Data":"b350ff55fd93c69942fd77ca393744433abc13f145b494dd68a72c35f873eee4"} Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.643178 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.706539 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt\") pod \"d635d90c-f19b-4646-bde8-bffe1a0da42a\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.706676 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage\") pod \"d635d90c-f19b-4646-bde8-bffe1a0da42a\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.706739 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb8gq\" (UniqueName: \"kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq\") pod \"d635d90c-f19b-4646-bde8-bffe1a0da42a\" (UID: \"d635d90c-f19b-4646-bde8-bffe1a0da42a\") " Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.706767 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "d635d90c-f19b-4646-bde8-bffe1a0da42a" (UID: "d635d90c-f19b-4646-bde8-bffe1a0da42a"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.710517 5018 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d635d90c-f19b-4646-bde8-bffe1a0da42a-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.712421 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq" (OuterVolumeSpecName: "kube-api-access-jb8gq") pod "d635d90c-f19b-4646-bde8-bffe1a0da42a" (UID: "d635d90c-f19b-4646-bde8-bffe1a0da42a"). InnerVolumeSpecName "kube-api-access-jb8gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.725317 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "d635d90c-f19b-4646-bde8-bffe1a0da42a" (UID: "d635d90c-f19b-4646-bde8-bffe1a0da42a"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.812081 5018 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d635d90c-f19b-4646-bde8-bffe1a0da42a-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:04 crc kubenswrapper[5018]: I1014 08:28:04.812122 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb8gq\" (UniqueName: \"kubernetes.io/projected/d635d90c-f19b-4646-bde8-bffe1a0da42a-kube-api-access-jb8gq\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:05 crc kubenswrapper[5018]: I1014 08:28:05.287033 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2mj5x" event={"ID":"d635d90c-f19b-4646-bde8-bffe1a0da42a","Type":"ContainerDied","Data":"7bc1ef92d49c205283824aae80735deb651b4cdbf7ed75f690948d303e83e8bb"} Oct 14 08:28:05 crc kubenswrapper[5018]: I1014 08:28:05.287084 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bc1ef92d49c205283824aae80735deb651b4cdbf7ed75f690948d303e83e8bb" Oct 14 08:28:05 crc kubenswrapper[5018]: I1014 08:28:05.287165 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2mj5x" Oct 14 08:28:05 crc kubenswrapper[5018]: I1014 08:28:05.959806 5018 scope.go:117] "RemoveContainer" containerID="65eb8594532e20620d327b9aea09bf441323d01f58be10e38d16566bc934e879" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.558873 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-2mj5x"] Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.565037 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-2mj5x"] Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.606252 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:28:06 crc kubenswrapper[5018]: E1014 08:28:06.606700 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.620487 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d635d90c-f19b-4646-bde8-bffe1a0da42a" path="/var/lib/kubelet/pods/d635d90c-f19b-4646-bde8-bffe1a0da42a/volumes" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.679741 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-5dvvt"] Oct 14 08:28:06 crc kubenswrapper[5018]: E1014 08:28:06.680073 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d635d90c-f19b-4646-bde8-bffe1a0da42a" containerName="storage" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.680088 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d635d90c-f19b-4646-bde8-bffe1a0da42a" containerName="storage" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.680208 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d635d90c-f19b-4646-bde8-bffe1a0da42a" containerName="storage" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.680662 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.683746 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.684122 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.684502 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.688519 5018 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5sfhc" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.713844 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5dvvt"] Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.739792 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.739894 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.739959 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6qwx\" (UniqueName: \"kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.841774 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.841908 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.841968 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6qwx\" (UniqueName: \"kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.842274 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.842750 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:06 crc kubenswrapper[5018]: I1014 08:28:06.869244 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6qwx\" (UniqueName: \"kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx\") pod \"crc-storage-crc-5dvvt\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:07 crc kubenswrapper[5018]: I1014 08:28:07.012928 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:07 crc kubenswrapper[5018]: I1014 08:28:07.330881 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5dvvt"] Oct 14 08:28:07 crc kubenswrapper[5018]: W1014 08:28:07.336856 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod021986de_e888_4860_99ba_af51d3e07f8f.slice/crio-d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b WatchSource:0}: Error finding container d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b: Status 404 returned error can't find the container with id d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b Oct 14 08:28:08 crc kubenswrapper[5018]: I1014 08:28:08.311490 5018 generic.go:334] "Generic (PLEG): container finished" podID="021986de-e888-4860-99ba-af51d3e07f8f" containerID="5cf0a3a686d336f1b9be44fba43e55fa976bbcf8ed9669fad7b9724775bb3c5c" exitCode=0 Oct 14 08:28:08 crc kubenswrapper[5018]: I1014 08:28:08.311685 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5dvvt" event={"ID":"021986de-e888-4860-99ba-af51d3e07f8f","Type":"ContainerDied","Data":"5cf0a3a686d336f1b9be44fba43e55fa976bbcf8ed9669fad7b9724775bb3c5c"} Oct 14 08:28:08 crc kubenswrapper[5018]: I1014 08:28:08.312048 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5dvvt" event={"ID":"021986de-e888-4860-99ba-af51d3e07f8f","Type":"ContainerStarted","Data":"d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b"} Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.697896 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.789027 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt\") pod \"021986de-e888-4860-99ba-af51d3e07f8f\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.789159 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage\") pod \"021986de-e888-4860-99ba-af51d3e07f8f\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.789188 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6qwx\" (UniqueName: \"kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx\") pod \"021986de-e888-4860-99ba-af51d3e07f8f\" (UID: \"021986de-e888-4860-99ba-af51d3e07f8f\") " Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.789191 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "021986de-e888-4860-99ba-af51d3e07f8f" (UID: "021986de-e888-4860-99ba-af51d3e07f8f"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.789507 5018 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021986de-e888-4860-99ba-af51d3e07f8f-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.796531 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx" (OuterVolumeSpecName: "kube-api-access-p6qwx") pod "021986de-e888-4860-99ba-af51d3e07f8f" (UID: "021986de-e888-4860-99ba-af51d3e07f8f"). InnerVolumeSpecName "kube-api-access-p6qwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.808095 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "021986de-e888-4860-99ba-af51d3e07f8f" (UID: "021986de-e888-4860-99ba-af51d3e07f8f"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.890594 5018 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021986de-e888-4860-99ba-af51d3e07f8f-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:09 crc kubenswrapper[5018]: I1014 08:28:09.890690 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6qwx\" (UniqueName: \"kubernetes.io/projected/021986de-e888-4860-99ba-af51d3e07f8f-kube-api-access-p6qwx\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:10 crc kubenswrapper[5018]: I1014 08:28:10.381544 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5dvvt" event={"ID":"021986de-e888-4860-99ba-af51d3e07f8f","Type":"ContainerDied","Data":"d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b"} Oct 14 08:28:10 crc kubenswrapper[5018]: I1014 08:28:10.381586 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8051de630b165308c5e66397157e78ae5a5396c182ed8a89e1e84c13361831b" Oct 14 08:28:10 crc kubenswrapper[5018]: I1014 08:28:10.381673 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5dvvt" Oct 14 08:28:21 crc kubenswrapper[5018]: I1014 08:28:21.605600 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:28:21 crc kubenswrapper[5018]: E1014 08:28:21.606607 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:28:34 crc kubenswrapper[5018]: I1014 08:28:34.617575 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:28:34 crc kubenswrapper[5018]: E1014 08:28:34.618758 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:28:46 crc kubenswrapper[5018]: I1014 08:28:46.606129 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:28:46 crc kubenswrapper[5018]: E1014 08:28:46.607095 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:28:59 crc kubenswrapper[5018]: I1014 08:28:59.605120 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:28:59 crc kubenswrapper[5018]: E1014 08:28:59.606698 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:29:14 crc kubenswrapper[5018]: I1014 08:29:14.613910 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:29:14 crc kubenswrapper[5018]: E1014 08:29:14.616405 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.907696 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:19 crc kubenswrapper[5018]: E1014 08:29:19.909136 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021986de-e888-4860-99ba-af51d3e07f8f" containerName="storage" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.909159 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="021986de-e888-4860-99ba-af51d3e07f8f" containerName="storage" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.909533 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="021986de-e888-4860-99ba-af51d3e07f8f" containerName="storage" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.911606 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.920464 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.962000 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.962204 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79zp\" (UniqueName: \"kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:19 crc kubenswrapper[5018]: I1014 08:29:19.962246 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.068219 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.068344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79zp\" (UniqueName: \"kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.068371 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.069002 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.069210 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.109032 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79zp\" (UniqueName: \"kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp\") pod \"certified-operators-27wld\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.252190 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:20 crc kubenswrapper[5018]: I1014 08:29:20.759394 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:21 crc kubenswrapper[5018]: I1014 08:29:21.088562 5018 generic.go:334] "Generic (PLEG): container finished" podID="298d5049-4608-42f4-9c25-809f78d86400" containerID="4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63" exitCode=0 Oct 14 08:29:21 crc kubenswrapper[5018]: I1014 08:29:21.088690 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerDied","Data":"4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63"} Oct 14 08:29:21 crc kubenswrapper[5018]: I1014 08:29:21.088753 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerStarted","Data":"2d0f49c05999067eb5b95409cff8acad3a94a0023baaa5c071e29e8639d3ceca"} Oct 14 08:29:21 crc kubenswrapper[5018]: I1014 08:29:21.090321 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:29:22 crc kubenswrapper[5018]: I1014 08:29:22.104250 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerStarted","Data":"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968"} Oct 14 08:29:23 crc kubenswrapper[5018]: I1014 08:29:23.117969 5018 generic.go:334] "Generic (PLEG): container finished" podID="298d5049-4608-42f4-9c25-809f78d86400" containerID="f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968" exitCode=0 Oct 14 08:29:23 crc kubenswrapper[5018]: I1014 08:29:23.118051 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerDied","Data":"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968"} Oct 14 08:29:24 crc kubenswrapper[5018]: I1014 08:29:24.132680 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerStarted","Data":"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a"} Oct 14 08:29:24 crc kubenswrapper[5018]: I1014 08:29:24.171606 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-27wld" podStartSLOduration=2.605284739 podStartE2EDuration="5.171579616s" podCreationTimestamp="2025-10-14 08:29:19 +0000 UTC" firstStartedPulling="2025-10-14 08:29:21.089978732 +0000 UTC m=+5977.674025369" lastFinishedPulling="2025-10-14 08:29:23.656273579 +0000 UTC m=+5980.240320246" observedRunningTime="2025-10-14 08:29:24.162198016 +0000 UTC m=+5980.746244653" watchObservedRunningTime="2025-10-14 08:29:24.171579616 +0000 UTC m=+5980.755626283" Oct 14 08:29:25 crc kubenswrapper[5018]: I1014 08:29:25.605179 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:29:25 crc kubenswrapper[5018]: E1014 08:29:25.605590 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:29:30 crc kubenswrapper[5018]: I1014 08:29:30.253003 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:30 crc kubenswrapper[5018]: I1014 08:29:30.254128 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:30 crc kubenswrapper[5018]: I1014 08:29:30.334205 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:31 crc kubenswrapper[5018]: I1014 08:29:31.256765 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:32 crc kubenswrapper[5018]: I1014 08:29:32.288208 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.227981 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-27wld" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="registry-server" containerID="cri-o://052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a" gracePeriod=2 Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.711164 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.732900 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities\") pod \"298d5049-4608-42f4-9c25-809f78d86400\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.733211 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content\") pod \"298d5049-4608-42f4-9c25-809f78d86400\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.733274 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j79zp\" (UniqueName: \"kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp\") pod \"298d5049-4608-42f4-9c25-809f78d86400\" (UID: \"298d5049-4608-42f4-9c25-809f78d86400\") " Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.735989 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities" (OuterVolumeSpecName: "utilities") pod "298d5049-4608-42f4-9c25-809f78d86400" (UID: "298d5049-4608-42f4-9c25-809f78d86400"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.742697 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp" (OuterVolumeSpecName: "kube-api-access-j79zp") pod "298d5049-4608-42f4-9c25-809f78d86400" (UID: "298d5049-4608-42f4-9c25-809f78d86400"). InnerVolumeSpecName "kube-api-access-j79zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.786346 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "298d5049-4608-42f4-9c25-809f78d86400" (UID: "298d5049-4608-42f4-9c25-809f78d86400"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.835097 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.835155 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/298d5049-4608-42f4-9c25-809f78d86400-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:34 crc kubenswrapper[5018]: I1014 08:29:34.835179 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j79zp\" (UniqueName: \"kubernetes.io/projected/298d5049-4608-42f4-9c25-809f78d86400-kube-api-access-j79zp\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.241476 5018 generic.go:334] "Generic (PLEG): container finished" podID="298d5049-4608-42f4-9c25-809f78d86400" containerID="052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a" exitCode=0 Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.241529 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerDied","Data":"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a"} Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.241594 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-27wld" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.241651 5018 scope.go:117] "RemoveContainer" containerID="052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.241600 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-27wld" event={"ID":"298d5049-4608-42f4-9c25-809f78d86400","Type":"ContainerDied","Data":"2d0f49c05999067eb5b95409cff8acad3a94a0023baaa5c071e29e8639d3ceca"} Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.280406 5018 scope.go:117] "RemoveContainer" containerID="f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.296665 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.305279 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-27wld"] Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.315528 5018 scope.go:117] "RemoveContainer" containerID="4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.364026 5018 scope.go:117] "RemoveContainer" containerID="052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a" Oct 14 08:29:35 crc kubenswrapper[5018]: E1014 08:29:35.364787 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a\": container with ID starting with 052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a not found: ID does not exist" containerID="052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.364867 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a"} err="failed to get container status \"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a\": rpc error: code = NotFound desc = could not find container \"052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a\": container with ID starting with 052184880eb2949e3d650afb627e9e65e0e8a67f2c6c30be9d191c261071a32a not found: ID does not exist" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.364915 5018 scope.go:117] "RemoveContainer" containerID="f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968" Oct 14 08:29:35 crc kubenswrapper[5018]: E1014 08:29:35.365660 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968\": container with ID starting with f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968 not found: ID does not exist" containerID="f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.365714 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968"} err="failed to get container status \"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968\": rpc error: code = NotFound desc = could not find container \"f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968\": container with ID starting with f719bf0fd95a301eba1577a6480eec23023dafec82ce4d47495b6645e9074968 not found: ID does not exist" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.365757 5018 scope.go:117] "RemoveContainer" containerID="4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63" Oct 14 08:29:35 crc kubenswrapper[5018]: E1014 08:29:35.366517 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63\": container with ID starting with 4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63 not found: ID does not exist" containerID="4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63" Oct 14 08:29:35 crc kubenswrapper[5018]: I1014 08:29:35.366567 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63"} err="failed to get container status \"4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63\": rpc error: code = NotFound desc = could not find container \"4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63\": container with ID starting with 4d713344a9af6a31081b6a33b95b6efdcb35f66a1f6fd58329012587bef51a63 not found: ID does not exist" Oct 14 08:29:36 crc kubenswrapper[5018]: I1014 08:29:36.621540 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298d5049-4608-42f4-9c25-809f78d86400" path="/var/lib/kubelet/pods/298d5049-4608-42f4-9c25-809f78d86400/volumes" Oct 14 08:29:37 crc kubenswrapper[5018]: I1014 08:29:37.605045 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:29:38 crc kubenswrapper[5018]: I1014 08:29:38.275532 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99"} Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.200082 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-57542"] Oct 14 08:30:00 crc kubenswrapper[5018]: E1014 08:30:00.201131 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="registry-server" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.201155 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="registry-server" Oct 14 08:30:00 crc kubenswrapper[5018]: E1014 08:30:00.201197 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="extract-content" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.201209 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="extract-content" Oct 14 08:30:00 crc kubenswrapper[5018]: E1014 08:30:00.201243 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="extract-utilities" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.201257 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="extract-utilities" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.201523 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="298d5049-4608-42f4-9c25-809f78d86400" containerName="registry-server" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.202317 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.205756 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.206105 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.215598 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-57542"] Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.268823 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.268937 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxmwt\" (UniqueName: \"kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.268986 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.370808 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.370950 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxmwt\" (UniqueName: \"kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.371007 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.373344 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.384844 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.396518 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxmwt\" (UniqueName: \"kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt\") pod \"collect-profiles-29340510-57542\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:00 crc kubenswrapper[5018]: I1014 08:30:00.554031 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:01 crc kubenswrapper[5018]: I1014 08:30:01.049439 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-57542"] Oct 14 08:30:01 crc kubenswrapper[5018]: W1014 08:30:01.067351 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod559ab8ea_2de3_4070_9ac2_24b3e548d8f4.slice/crio-6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd WatchSource:0}: Error finding container 6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd: Status 404 returned error can't find the container with id 6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd Oct 14 08:30:01 crc kubenswrapper[5018]: I1014 08:30:01.504038 5018 generic.go:334] "Generic (PLEG): container finished" podID="559ab8ea-2de3-4070-9ac2-24b3e548d8f4" containerID="c7d082c5e6109a62ef8b9e7a9e4006319408c065e9bdf5f41aeedf98201ffee8" exitCode=0 Oct 14 08:30:01 crc kubenswrapper[5018]: I1014 08:30:01.504103 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" event={"ID":"559ab8ea-2de3-4070-9ac2-24b3e548d8f4","Type":"ContainerDied","Data":"c7d082c5e6109a62ef8b9e7a9e4006319408c065e9bdf5f41aeedf98201ffee8"} Oct 14 08:30:01 crc kubenswrapper[5018]: I1014 08:30:01.504503 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" event={"ID":"559ab8ea-2de3-4070-9ac2-24b3e548d8f4","Type":"ContainerStarted","Data":"6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd"} Oct 14 08:30:02 crc kubenswrapper[5018]: I1014 08:30:02.949723 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.123907 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume\") pod \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.124025 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxmwt\" (UniqueName: \"kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt\") pod \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.124198 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume\") pod \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\" (UID: \"559ab8ea-2de3-4070-9ac2-24b3e548d8f4\") " Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.124876 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume" (OuterVolumeSpecName: "config-volume") pod "559ab8ea-2de3-4070-9ac2-24b3e548d8f4" (UID: "559ab8ea-2de3-4070-9ac2-24b3e548d8f4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.133306 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "559ab8ea-2de3-4070-9ac2-24b3e548d8f4" (UID: "559ab8ea-2de3-4070-9ac2-24b3e548d8f4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.133397 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt" (OuterVolumeSpecName: "kube-api-access-pxmwt") pod "559ab8ea-2de3-4070-9ac2-24b3e548d8f4" (UID: "559ab8ea-2de3-4070-9ac2-24b3e548d8f4"). InnerVolumeSpecName "kube-api-access-pxmwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.225670 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxmwt\" (UniqueName: \"kubernetes.io/projected/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-kube-api-access-pxmwt\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.225710 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.225722 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/559ab8ea-2de3-4070-9ac2-24b3e548d8f4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.537903 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" event={"ID":"559ab8ea-2de3-4070-9ac2-24b3e548d8f4","Type":"ContainerDied","Data":"6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd"} Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.537961 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d410b2ec904987005786d08e928427dfe226d766b1556a71d02ce6d3a4843cd" Oct 14 08:30:03 crc kubenswrapper[5018]: I1014 08:30:03.538665 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-57542" Oct 14 08:30:04 crc kubenswrapper[5018]: I1014 08:30:04.002213 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk"] Oct 14 08:30:04 crc kubenswrapper[5018]: I1014 08:30:04.011066 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-mmxdk"] Oct 14 08:30:04 crc kubenswrapper[5018]: I1014 08:30:04.620135 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a042892-04be-4f31-8f45-db623910ed9e" path="/var/lib/kubelet/pods/8a042892-04be-4f31-8f45-db623910ed9e/volumes" Oct 14 08:30:06 crc kubenswrapper[5018]: I1014 08:30:06.070428 5018 scope.go:117] "RemoveContainer" containerID="6214bd48018768a8aba58b50dba689a5cdd48223db732fb3a9a50fd68a65d538" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.217446 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:20 crc kubenswrapper[5018]: E1014 08:30:20.218253 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559ab8ea-2de3-4070-9ac2-24b3e548d8f4" containerName="collect-profiles" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.218266 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="559ab8ea-2de3-4070-9ac2-24b3e548d8f4" containerName="collect-profiles" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.218386 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="559ab8ea-2de3-4070-9ac2-24b3e548d8f4" containerName="collect-profiles" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.219208 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.220317 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.223364 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.224096 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.224279 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-mhxw6" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.224469 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.290539 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.298959 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.303759 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.303918 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.324438 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qj8\" (UniqueName: \"kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.324487 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.426024 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td7bs\" (UniqueName: \"kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.426075 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.426120 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qj8\" (UniqueName: \"kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.426138 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.426160 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.427291 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.456020 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qj8\" (UniqueName: \"kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8\") pod \"dnsmasq-dns-dbc6498b5-5vtgg\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.528989 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td7bs\" (UniqueName: \"kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.529049 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.529122 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.529918 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.529927 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.537170 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.552361 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td7bs\" (UniqueName: \"kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs\") pod \"dnsmasq-dns-55d5b9f44c-htb7f\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.559206 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.560303 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.615384 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.617026 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.648243 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.733028 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.733153 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.733182 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2r9n\" (UniqueName: \"kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.834244 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.834402 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.834431 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2r9n\" (UniqueName: \"kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.835324 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.835915 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.853274 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2r9n\" (UniqueName: \"kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n\") pod \"dnsmasq-dns-5c658747f5-tbpn9\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.875251 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:20 crc kubenswrapper[5018]: I1014 08:30:20.947066 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.052727 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.065906 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.099405 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.101426 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.128940 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.241380 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.241530 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5xfn\" (UniqueName: \"kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.241552 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.342742 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5xfn\" (UniqueName: \"kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.342787 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.342850 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.343600 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.343701 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.359373 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5xfn\" (UniqueName: \"kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn\") pod \"dnsmasq-dns-54bcdcd99f-qtkqs\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.429285 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.455328 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.720173 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" event={"ID":"e36da379-27eb-4f3f-b496-7e9658dfed07","Type":"ContainerStarted","Data":"81b6b7a51a2fed6d425cfa86fec5eae2b256f49b6a99360a3c886227b8b6d6ad"} Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.722470 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" event={"ID":"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b","Type":"ContainerStarted","Data":"ebb0d81cb8c1ec87aa092ef2ca939e8d7c0120d0722ff28e26815da2b19d1753"} Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.730162 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" event={"ID":"cecfd985-568e-4bc2-912c-4d1363ceb686","Type":"ContainerStarted","Data":"dedd5a6e29d44db00e4083dcb7a6dee46c7e8501895b5dc448d2cf16dcf02778"} Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.740957 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.743538 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.745640 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746421 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746554 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746745 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746802 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746953 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4qpkf" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.746812 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.757409 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.861676 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862261 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862320 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862347 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862393 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862421 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862441 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862459 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862474 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbgh\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862500 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.862521 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: W1014 08:30:21.869677 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd11183f_f4b4_4c8d_9eaf_860263e0fb21.slice/crio-5d9f0d39bcab8ea4a3166aa17924fa22d96d9f9598a2643e953e63d1903bb862 WatchSource:0}: Error finding container 5d9f0d39bcab8ea4a3166aa17924fa22d96d9f9598a2643e953e63d1903bb862: Status 404 returned error can't find the container with id 5d9f0d39bcab8ea4a3166aa17924fa22d96d9f9598a2643e953e63d1903bb862 Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.963916 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.963974 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.963997 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbgh\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964015 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964041 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964064 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964095 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964124 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964146 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964181 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964203 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.964663 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.965172 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.965664 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.965830 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.966394 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.968075 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.968108 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/927488b0368b94c1e5e36dc99b7c109dfb844e110a1f467b4cc605c2894a8706/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.971369 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.976800 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.979469 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:21 crc kubenswrapper[5018]: I1014 08:30:21.992234 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbgh\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.003599 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.064211 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.077638 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.231283 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.232673 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.236309 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.237868 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.237998 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8c9vp" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.238117 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.238220 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.238351 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.238542 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.251233 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370445 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370494 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370522 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370545 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370568 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370587 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370612 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6kxd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370687 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370704 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370728 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.370748 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471559 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6kxd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471651 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471675 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471703 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471726 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471857 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471877 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471902 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471927 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471949 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.471969 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.472266 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.472933 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.472979 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.473774 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.474376 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.476298 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.476321 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/415a1d7451e47b1410263ebc620512fbbd328bf639d64d5b904d6662fd2628ee/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.477346 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.477694 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.477786 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.478164 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.487804 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6kxd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.504631 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.590507 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.637819 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.744335 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" event={"ID":"bd11183f-f4b4-4c8d-9eaf-860263e0fb21","Type":"ContainerStarted","Data":"5d9f0d39bcab8ea4a3166aa17924fa22d96d9f9598a2643e953e63d1903bb862"} Oct 14 08:30:22 crc kubenswrapper[5018]: I1014 08:30:22.745292 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerStarted","Data":"7925c82d00af014ecaac9bc8d3bd931652f5b350ecf6089477413971ba9b14fa"} Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.000371 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.001775 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.007126 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.009303 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.010353 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-zqpsg" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.010513 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.010649 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.013614 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.016031 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.048467 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:30:23 crc kubenswrapper[5018]: W1014 08:30:23.055458 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fd9eca2_b576_4f12_a91e_df9042db3c44.slice/crio-5f2c0a7c8618d3843e388d3fbe79a3bdfccd99fa6d1eb715215d662a368fd797 WatchSource:0}: Error finding container 5f2c0a7c8618d3843e388d3fbe79a3bdfccd99fa6d1eb715215d662a368fd797: Status 404 returned error can't find the container with id 5f2c0a7c8618d3843e388d3fbe79a3bdfccd99fa6d1eb715215d662a368fd797 Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.100513 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.100598 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.100766 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.100836 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.101006 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzq5c\" (UniqueName: \"kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.101065 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.101166 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.101204 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.101300 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202581 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202670 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202694 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202722 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202766 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202784 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202819 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202839 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.202866 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzq5c\" (UniqueName: \"kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.203719 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.204736 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.205758 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.206232 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.206284 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.206315 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ba4d3d49b757e3d8f2351ffd51c0b5b5da0e8cf487cbeedcc79d8b18e664f0c3/globalmount\"" pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.209356 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.209509 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.210556 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.218388 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzq5c\" (UniqueName: \"kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.236334 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") pod \"openstack-galera-0\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.343215 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.756551 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerStarted","Data":"5f2c0a7c8618d3843e388d3fbe79a3bdfccd99fa6d1eb715215d662a368fd797"} Oct 14 08:30:23 crc kubenswrapper[5018]: I1014 08:30:23.844333 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.376162 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.378477 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.382933 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.383088 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.383210 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-g8wtx" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.383436 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.409188 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524238 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524293 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524341 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524365 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524588 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524630 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524733 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.524764 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nljm8\" (UniqueName: \"kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626038 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626086 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nljm8\" (UniqueName: \"kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626130 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626172 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626232 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626272 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.626319 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.627255 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.629325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.629651 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.629689 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/38de2f563aab74d86562106de977082d294a196844c0c3574cd9e3eae560a2ca/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.630443 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.631292 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.642329 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.648921 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.649231 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nljm8\" (UniqueName: \"kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.650139 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.681777 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") pod \"openstack-cell1-galera-0\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.718674 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.768268 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerStarted","Data":"5bd6d3322f32483a18d6f7509f368d72cea2754b64b07b54604a283d01808f05"} Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.836645 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.840798 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.845011 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.845011 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.845134 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-r52bz" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.852965 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.945265 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.945321 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.945347 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.945368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt2lf\" (UniqueName: \"kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:24 crc kubenswrapper[5018]: I1014 08:30:24.945407 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.046774 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.046900 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.046931 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.046955 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.046977 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt2lf\" (UniqueName: \"kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.048143 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.048208 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.058092 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.058462 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.062418 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt2lf\" (UniqueName: \"kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf\") pod \"memcached-0\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.156302 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.261082 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:30:25 crc kubenswrapper[5018]: W1014 08:30:25.274068 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a5a6c5b_9410_4fde_98c4_112a882b366c.slice/crio-557b6127928f5f5395cc76a2bae8881ba30a62f5b03767fefc232ab4759c7403 WatchSource:0}: Error finding container 557b6127928f5f5395cc76a2bae8881ba30a62f5b03767fefc232ab4759c7403: Status 404 returned error can't find the container with id 557b6127928f5f5395cc76a2bae8881ba30a62f5b03767fefc232ab4759c7403 Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.587329 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.778285 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"941d01f1-c45d-49c3-923b-d685ad3a2f4b","Type":"ContainerStarted","Data":"85ec1a5eb8758d65e8e0f323dfef6a54fe81f358c97e49d594c9364bb909b70c"} Oct 14 08:30:25 crc kubenswrapper[5018]: I1014 08:30:25.779647 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerStarted","Data":"557b6127928f5f5395cc76a2bae8881ba30a62f5b03767fefc232ab4759c7403"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.890231 5018 generic.go:334] "Generic (PLEG): container finished" podID="54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" containerID="51ee8ca04044d5726040fefb26ef01eb2d72bd10ecfa7e775c6055245bc0f20c" exitCode=0 Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.890347 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" event={"ID":"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b","Type":"ContainerDied","Data":"51ee8ca04044d5726040fefb26ef01eb2d72bd10ecfa7e775c6055245bc0f20c"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.895680 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerStarted","Data":"97f38af704283bb352ed02f09332ddce7ba2726b12ca8ffa9339d06c33f07de1"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.898072 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerStarted","Data":"85f4670aa47a47b14a062924c76d871821892d2a11c13336e362bd97129914b3"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.900280 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"941d01f1-c45d-49c3-923b-d685ad3a2f4b","Type":"ContainerStarted","Data":"49e3187ba65a221c522e49e4c2118872d506a95aa0ad3018a57afeb0f6482e51"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.900472 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.901723 5018 generic.go:334] "Generic (PLEG): container finished" podID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerID="1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350" exitCode=0 Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.901778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" event={"ID":"cecfd985-568e-4bc2-912c-4d1363ceb686","Type":"ContainerDied","Data":"1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.904994 5018 generic.go:334] "Generic (PLEG): container finished" podID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerID="f0b13b23065addf5b99e5a14ff66154b7d69fff3f698ea5629035c29186c1000" exitCode=0 Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.905067 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" event={"ID":"bd11183f-f4b4-4c8d-9eaf-860263e0fb21","Type":"ContainerDied","Data":"f0b13b23065addf5b99e5a14ff66154b7d69fff3f698ea5629035c29186c1000"} Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.911884 5018 generic.go:334] "Generic (PLEG): container finished" podID="e36da379-27eb-4f3f-b496-7e9658dfed07" containerID="12d104428ca7b649dbb373c4ef3abdd766d2180328fa4a80f24637a32ddcea9b" exitCode=0 Oct 14 08:30:38 crc kubenswrapper[5018]: I1014 08:30:38.911949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" event={"ID":"e36da379-27eb-4f3f-b496-7e9658dfed07","Type":"ContainerDied","Data":"12d104428ca7b649dbb373c4ef3abdd766d2180328fa4a80f24637a32ddcea9b"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.046322 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.351385858 podStartE2EDuration="15.046306176s" podCreationTimestamp="2025-10-14 08:30:24 +0000 UTC" firstStartedPulling="2025-10-14 08:30:25.598038903 +0000 UTC m=+6042.182085530" lastFinishedPulling="2025-10-14 08:30:38.292959221 +0000 UTC m=+6054.877005848" observedRunningTime="2025-10-14 08:30:39.041696864 +0000 UTC m=+6055.625743501" watchObservedRunningTime="2025-10-14 08:30:39.046306176 +0000 UTC m=+6055.630352803" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.255506 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.395558 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc\") pod \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.395736 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config\") pod \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.395920 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td7bs\" (UniqueName: \"kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs\") pod \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\" (UID: \"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.400532 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs" (OuterVolumeSpecName: "kube-api-access-td7bs") pod "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" (UID: "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b"). InnerVolumeSpecName "kube-api-access-td7bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.432468 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" (UID: "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.499183 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td7bs\" (UniqueName: \"kubernetes.io/projected/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-kube-api-access-td7bs\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.499259 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.516036 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config" (OuterVolumeSpecName: "config") pod "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" (UID: "54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.601540 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.614893 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.704278 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc\") pod \"e36da379-27eb-4f3f-b496-7e9658dfed07\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.704346 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2r9n\" (UniqueName: \"kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n\") pod \"e36da379-27eb-4f3f-b496-7e9658dfed07\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.704605 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config\") pod \"e36da379-27eb-4f3f-b496-7e9658dfed07\" (UID: \"e36da379-27eb-4f3f-b496-7e9658dfed07\") " Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.782324 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n" (OuterVolumeSpecName: "kube-api-access-b2r9n") pod "e36da379-27eb-4f3f-b496-7e9658dfed07" (UID: "e36da379-27eb-4f3f-b496-7e9658dfed07"). InnerVolumeSpecName "kube-api-access-b2r9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.783989 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config" (OuterVolumeSpecName: "config") pod "e36da379-27eb-4f3f-b496-7e9658dfed07" (UID: "e36da379-27eb-4f3f-b496-7e9658dfed07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.787124 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e36da379-27eb-4f3f-b496-7e9658dfed07" (UID: "e36da379-27eb-4f3f-b496-7e9658dfed07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.807325 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.807358 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e36da379-27eb-4f3f-b496-7e9658dfed07-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.807369 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2r9n\" (UniqueName: \"kubernetes.io/projected/e36da379-27eb-4f3f-b496-7e9658dfed07-kube-api-access-b2r9n\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.918976 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" event={"ID":"bd11183f-f4b4-4c8d-9eaf-860263e0fb21","Type":"ContainerStarted","Data":"72a2313ccdc343aa69947575857f12b150ba17b3bead5ec4734ffa7051a61ad8"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.919043 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.921544 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerStarted","Data":"652b7b782181f73ef165e88fbce3123b71544d2476428f611fc5be1e48289969"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.923154 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" event={"ID":"e36da379-27eb-4f3f-b496-7e9658dfed07","Type":"ContainerDied","Data":"81b6b7a51a2fed6d425cfa86fec5eae2b256f49b6a99360a3c886227b8b6d6ad"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.923185 5018 scope.go:117] "RemoveContainer" containerID="12d104428ca7b649dbb373c4ef3abdd766d2180328fa4a80f24637a32ddcea9b" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.923268 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c658747f5-tbpn9" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.933868 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.933871 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d5b9f44c-htb7f" event={"ID":"54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b","Type":"ContainerDied","Data":"ebb0d81cb8c1ec87aa092ef2ca939e8d7c0120d0722ff28e26815da2b19d1753"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.939825 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerStarted","Data":"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.945014 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" event={"ID":"cecfd985-568e-4bc2-912c-4d1363ceb686","Type":"ContainerStarted","Data":"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30"} Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.945375 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.946953 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" podStartSLOduration=2.507245139 podStartE2EDuration="18.946928763s" podCreationTimestamp="2025-10-14 08:30:21 +0000 UTC" firstStartedPulling="2025-10-14 08:30:21.872509089 +0000 UTC m=+6038.456555706" lastFinishedPulling="2025-10-14 08:30:38.312192703 +0000 UTC m=+6054.896239330" observedRunningTime="2025-10-14 08:30:39.937731789 +0000 UTC m=+6056.521778416" watchObservedRunningTime="2025-10-14 08:30:39.946928763 +0000 UTC m=+6056.530975420" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.960975 5018 scope.go:117] "RemoveContainer" containerID="51ee8ca04044d5726040fefb26ef01eb2d72bd10ecfa7e775c6055245bc0f20c" Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.982759 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:39 crc kubenswrapper[5018]: I1014 08:30:39.989554 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c658747f5-tbpn9"] Oct 14 08:30:40 crc kubenswrapper[5018]: I1014 08:30:40.029883 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" podStartSLOduration=2.721450441 podStartE2EDuration="20.029862766s" podCreationTimestamp="2025-10-14 08:30:20 +0000 UTC" firstStartedPulling="2025-10-14 08:30:21.068020285 +0000 UTC m=+6037.652066912" lastFinishedPulling="2025-10-14 08:30:38.37643261 +0000 UTC m=+6054.960479237" observedRunningTime="2025-10-14 08:30:40.021880366 +0000 UTC m=+6056.605927003" watchObservedRunningTime="2025-10-14 08:30:40.029862766 +0000 UTC m=+6056.613909403" Oct 14 08:30:40 crc kubenswrapper[5018]: I1014 08:30:40.083917 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:40 crc kubenswrapper[5018]: I1014 08:30:40.084381 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55d5b9f44c-htb7f"] Oct 14 08:30:40 crc kubenswrapper[5018]: I1014 08:30:40.626032 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" path="/var/lib/kubelet/pods/54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b/volumes" Oct 14 08:30:40 crc kubenswrapper[5018]: I1014 08:30:40.626822 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36da379-27eb-4f3f-b496-7e9658dfed07" path="/var/lib/kubelet/pods/e36da379-27eb-4f3f-b496-7e9658dfed07/volumes" Oct 14 08:30:42 crc kubenswrapper[5018]: I1014 08:30:42.978403 5018 generic.go:334] "Generic (PLEG): container finished" podID="1383035e-278d-425c-9062-f76183fcdb31" containerID="97f38af704283bb352ed02f09332ddce7ba2726b12ca8ffa9339d06c33f07de1" exitCode=0 Oct 14 08:30:42 crc kubenswrapper[5018]: I1014 08:30:42.978679 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerDied","Data":"97f38af704283bb352ed02f09332ddce7ba2726b12ca8ffa9339d06c33f07de1"} Oct 14 08:30:42 crc kubenswrapper[5018]: I1014 08:30:42.984891 5018 generic.go:334] "Generic (PLEG): container finished" podID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerID="85f4670aa47a47b14a062924c76d871821892d2a11c13336e362bd97129914b3" exitCode=0 Oct 14 08:30:42 crc kubenswrapper[5018]: I1014 08:30:42.984962 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerDied","Data":"85f4670aa47a47b14a062924c76d871821892d2a11c13336e362bd97129914b3"} Oct 14 08:30:43 crc kubenswrapper[5018]: I1014 08:30:43.999351 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerStarted","Data":"589d1bdd70b59fd98fe13f906d2c0e56fa29052d702a35f68536bb9f4b51c629"} Oct 14 08:30:44 crc kubenswrapper[5018]: I1014 08:30:44.003277 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerStarted","Data":"005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711"} Oct 14 08:30:44 crc kubenswrapper[5018]: I1014 08:30:44.043405 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.062736215 podStartE2EDuration="23.043372664s" podCreationTimestamp="2025-10-14 08:30:21 +0000 UTC" firstStartedPulling="2025-10-14 08:30:24.342321874 +0000 UTC m=+6040.926368501" lastFinishedPulling="2025-10-14 08:30:38.322958323 +0000 UTC m=+6054.907004950" observedRunningTime="2025-10-14 08:30:44.030317019 +0000 UTC m=+6060.614363656" watchObservedRunningTime="2025-10-14 08:30:44.043372664 +0000 UTC m=+6060.627419331" Oct 14 08:30:44 crc kubenswrapper[5018]: I1014 08:30:44.083437 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.984931651 podStartE2EDuration="21.083385474s" podCreationTimestamp="2025-10-14 08:30:23 +0000 UTC" firstStartedPulling="2025-10-14 08:30:25.278196513 +0000 UTC m=+6041.862243180" lastFinishedPulling="2025-10-14 08:30:38.376650366 +0000 UTC m=+6054.960697003" observedRunningTime="2025-10-14 08:30:44.074595541 +0000 UTC m=+6060.658642248" watchObservedRunningTime="2025-10-14 08:30:44.083385474 +0000 UTC m=+6060.667432121" Oct 14 08:30:44 crc kubenswrapper[5018]: I1014 08:30:44.724143 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:44 crc kubenswrapper[5018]: I1014 08:30:44.724190 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:45 crc kubenswrapper[5018]: I1014 08:30:45.157964 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 08:30:45 crc kubenswrapper[5018]: I1014 08:30:45.539886 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:46 crc kubenswrapper[5018]: I1014 08:30:46.431535 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:30:46 crc kubenswrapper[5018]: I1014 08:30:46.501130 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:46 crc kubenswrapper[5018]: I1014 08:30:46.501365 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="dnsmasq-dns" containerID="cri-o://78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30" gracePeriod=10 Oct 14 08:30:46 crc kubenswrapper[5018]: I1014 08:30:46.995283 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.028463 5018 generic.go:334] "Generic (PLEG): container finished" podID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerID="78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30" exitCode=0 Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.028539 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.028547 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" event={"ID":"cecfd985-568e-4bc2-912c-4d1363ceb686","Type":"ContainerDied","Data":"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30"} Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.028942 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbc6498b5-5vtgg" event={"ID":"cecfd985-568e-4bc2-912c-4d1363ceb686","Type":"ContainerDied","Data":"dedd5a6e29d44db00e4083dcb7a6dee46c7e8501895b5dc448d2cf16dcf02778"} Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.028977 5018 scope.go:117] "RemoveContainer" containerID="78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.049169 5018 scope.go:117] "RemoveContainer" containerID="1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.067359 5018 scope.go:117] "RemoveContainer" containerID="78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30" Oct 14 08:30:47 crc kubenswrapper[5018]: E1014 08:30:47.068076 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30\": container with ID starting with 78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30 not found: ID does not exist" containerID="78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.068107 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30"} err="failed to get container status \"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30\": rpc error: code = NotFound desc = could not find container \"78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30\": container with ID starting with 78e50229106ffacd6db9f378def955dedf296abc4237bbc5210c12e44adf7b30 not found: ID does not exist" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.068143 5018 scope.go:117] "RemoveContainer" containerID="1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350" Oct 14 08:30:47 crc kubenswrapper[5018]: E1014 08:30:47.068459 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350\": container with ID starting with 1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350 not found: ID does not exist" containerID="1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.068550 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350"} err="failed to get container status \"1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350\": rpc error: code = NotFound desc = could not find container \"1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350\": container with ID starting with 1e92540e88e3a5c6f086051f001044d722f3a37296fae0d62f10eb408e66b350 not found: ID does not exist" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.141310 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qj8\" (UniqueName: \"kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8\") pod \"cecfd985-568e-4bc2-912c-4d1363ceb686\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.141377 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config\") pod \"cecfd985-568e-4bc2-912c-4d1363ceb686\" (UID: \"cecfd985-568e-4bc2-912c-4d1363ceb686\") " Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.148955 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8" (OuterVolumeSpecName: "kube-api-access-n7qj8") pod "cecfd985-568e-4bc2-912c-4d1363ceb686" (UID: "cecfd985-568e-4bc2-912c-4d1363ceb686"). InnerVolumeSpecName "kube-api-access-n7qj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.201360 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config" (OuterVolumeSpecName: "config") pod "cecfd985-568e-4bc2-912c-4d1363ceb686" (UID: "cecfd985-568e-4bc2-912c-4d1363ceb686"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.243601 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cecfd985-568e-4bc2-912c-4d1363ceb686-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.243673 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qj8\" (UniqueName: \"kubernetes.io/projected/cecfd985-568e-4bc2-912c-4d1363ceb686-kube-api-access-n7qj8\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.373134 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:47 crc kubenswrapper[5018]: I1014 08:30:47.379257 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dbc6498b5-5vtgg"] Oct 14 08:30:48 crc kubenswrapper[5018]: I1014 08:30:48.623340 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" path="/var/lib/kubelet/pods/cecfd985-568e-4bc2-912c-4d1363ceb686/volumes" Oct 14 08:30:48 crc kubenswrapper[5018]: I1014 08:30:48.812124 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:48 crc kubenswrapper[5018]: I1014 08:30:48.904779 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 08:30:53 crc kubenswrapper[5018]: I1014 08:30:53.344027 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 08:30:53 crc kubenswrapper[5018]: I1014 08:30:53.344477 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 08:30:53 crc kubenswrapper[5018]: I1014 08:30:53.436612 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 08:30:54 crc kubenswrapper[5018]: I1014 08:30:54.163215 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 08:31:12 crc kubenswrapper[5018]: E1014 08:31:12.457778 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod090c4f30_6438_4cd0_ad4d_8ab6531cfce5.slice/crio-conmon-a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:31:13 crc kubenswrapper[5018]: I1014 08:31:13.287446 5018 generic.go:334] "Generic (PLEG): container finished" podID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerID="652b7b782181f73ef165e88fbce3123b71544d2476428f611fc5be1e48289969" exitCode=0 Oct 14 08:31:13 crc kubenswrapper[5018]: I1014 08:31:13.287557 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerDied","Data":"652b7b782181f73ef165e88fbce3123b71544d2476428f611fc5be1e48289969"} Oct 14 08:31:13 crc kubenswrapper[5018]: I1014 08:31:13.289872 5018 generic.go:334] "Generic (PLEG): container finished" podID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerID="a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a" exitCode=0 Oct 14 08:31:13 crc kubenswrapper[5018]: I1014 08:31:13.289925 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerDied","Data":"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a"} Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.308499 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerStarted","Data":"fb861d8c016bc76027416920bf928d0a18a84db9ee084f7cac20073cef09ce36"} Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.309226 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.312130 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerStarted","Data":"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b"} Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.313534 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.341690 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.067914028 podStartE2EDuration="53.341661711s" podCreationTimestamp="2025-10-14 08:30:21 +0000 UTC" firstStartedPulling="2025-10-14 08:30:23.057509008 +0000 UTC m=+6039.641555635" lastFinishedPulling="2025-10-14 08:30:38.331256681 +0000 UTC m=+6054.915303318" observedRunningTime="2025-10-14 08:31:14.336914445 +0000 UTC m=+6090.920961092" watchObservedRunningTime="2025-10-14 08:31:14.341661711 +0000 UTC m=+6090.925708348" Oct 14 08:31:14 crc kubenswrapper[5018]: I1014 08:31:14.374874 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.74957444 podStartE2EDuration="54.374855905s" podCreationTimestamp="2025-10-14 08:30:20 +0000 UTC" firstStartedPulling="2025-10-14 08:30:22.662866688 +0000 UTC m=+6039.246913315" lastFinishedPulling="2025-10-14 08:30:38.288148133 +0000 UTC m=+6054.872194780" observedRunningTime="2025-10-14 08:31:14.36807038 +0000 UTC m=+6090.952117047" watchObservedRunningTime="2025-10-14 08:31:14.374855905 +0000 UTC m=+6090.958902532" Oct 14 08:31:32 crc kubenswrapper[5018]: I1014 08:31:32.082268 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:32 crc kubenswrapper[5018]: I1014 08:31:32.594955 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.991298 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:31:37 crc kubenswrapper[5018]: E1014 08:31:37.993314 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993355 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: E1014 08:31:37.993379 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36da379-27eb-4f3f-b496-7e9658dfed07" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993392 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36da379-27eb-4f3f-b496-7e9658dfed07" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: E1014 08:31:37.993447 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="dnsmasq-dns" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993487 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="dnsmasq-dns" Oct 14 08:31:37 crc kubenswrapper[5018]: E1014 08:31:37.993505 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993517 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993793 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36da379-27eb-4f3f-b496-7e9658dfed07" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993811 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="54b01dda-6cd3-436e-9ffe-7ca5f8e4dd4b" containerName="init" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.993849 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cecfd985-568e-4bc2-912c-4d1363ceb686" containerName="dnsmasq-dns" Oct 14 08:31:37 crc kubenswrapper[5018]: I1014 08:31:37.995195 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.010264 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.060920 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljwmr\" (UniqueName: \"kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.061084 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.061229 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.163112 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.163174 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljwmr\" (UniqueName: \"kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.163246 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.164370 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.164413 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.187563 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljwmr\" (UniqueName: \"kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr\") pod \"dnsmasq-dns-7ccc84877c-kwvdx\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.327452 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.674634 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:38 crc kubenswrapper[5018]: I1014 08:31:38.867201 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:31:38 crc kubenswrapper[5018]: W1014 08:31:38.882333 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9383f078_40c5_48ab_937b_31593364ecc7.slice/crio-333bca4329bdd793ae9b73e23e3b7dec13a9034a2c971d43dc1196a7bd03f63c WatchSource:0}: Error finding container 333bca4329bdd793ae9b73e23e3b7dec13a9034a2c971d43dc1196a7bd03f63c: Status 404 returned error can't find the container with id 333bca4329bdd793ae9b73e23e3b7dec13a9034a2c971d43dc1196a7bd03f63c Oct 14 08:31:39 crc kubenswrapper[5018]: I1014 08:31:39.443799 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:39 crc kubenswrapper[5018]: I1014 08:31:39.548652 5018 generic.go:334] "Generic (PLEG): container finished" podID="9383f078-40c5-48ab-937b-31593364ecc7" containerID="e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d" exitCode=0 Oct 14 08:31:39 crc kubenswrapper[5018]: I1014 08:31:39.548704 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" event={"ID":"9383f078-40c5-48ab-937b-31593364ecc7","Type":"ContainerDied","Data":"e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d"} Oct 14 08:31:39 crc kubenswrapper[5018]: I1014 08:31:39.548736 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" event={"ID":"9383f078-40c5-48ab-937b-31593364ecc7","Type":"ContainerStarted","Data":"333bca4329bdd793ae9b73e23e3b7dec13a9034a2c971d43dc1196a7bd03f63c"} Oct 14 08:31:40 crc kubenswrapper[5018]: I1014 08:31:40.559242 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" event={"ID":"9383f078-40c5-48ab-937b-31593364ecc7","Type":"ContainerStarted","Data":"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde"} Oct 14 08:31:40 crc kubenswrapper[5018]: I1014 08:31:40.559396 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:40 crc kubenswrapper[5018]: I1014 08:31:40.589318 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" podStartSLOduration=3.589299351 podStartE2EDuration="3.589299351s" podCreationTimestamp="2025-10-14 08:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:31:40.583416232 +0000 UTC m=+6117.167462869" watchObservedRunningTime="2025-10-14 08:31:40.589299351 +0000 UTC m=+6117.173345988" Oct 14 08:31:43 crc kubenswrapper[5018]: I1014 08:31:43.117954 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="rabbitmq" containerID="cri-o://fb861d8c016bc76027416920bf928d0a18a84db9ee084f7cac20073cef09ce36" gracePeriod=604796 Oct 14 08:31:43 crc kubenswrapper[5018]: I1014 08:31:43.738502 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="rabbitmq" containerID="cri-o://f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b" gracePeriod=604796 Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.328947 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.415448 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.415843 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="dnsmasq-dns" containerID="cri-o://72a2313ccdc343aa69947575857f12b150ba17b3bead5ec4734ffa7051a61ad8" gracePeriod=10 Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.678151 5018 generic.go:334] "Generic (PLEG): container finished" podID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerID="72a2313ccdc343aa69947575857f12b150ba17b3bead5ec4734ffa7051a61ad8" exitCode=0 Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.678342 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" event={"ID":"bd11183f-f4b4-4c8d-9eaf-860263e0fb21","Type":"ContainerDied","Data":"72a2313ccdc343aa69947575857f12b150ba17b3bead5ec4734ffa7051a61ad8"} Oct 14 08:31:48 crc kubenswrapper[5018]: I1014 08:31:48.871895 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.069734 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5xfn\" (UniqueName: \"kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn\") pod \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.069894 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config\") pod \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.069961 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc\") pod \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\" (UID: \"bd11183f-f4b4-4c8d-9eaf-860263e0fb21\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.080155 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn" (OuterVolumeSpecName: "kube-api-access-x5xfn") pod "bd11183f-f4b4-4c8d-9eaf-860263e0fb21" (UID: "bd11183f-f4b4-4c8d-9eaf-860263e0fb21"). InnerVolumeSpecName "kube-api-access-x5xfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.145823 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config" (OuterVolumeSpecName: "config") pod "bd11183f-f4b4-4c8d-9eaf-860263e0fb21" (UID: "bd11183f-f4b4-4c8d-9eaf-860263e0fb21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.147037 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd11183f-f4b4-4c8d-9eaf-860263e0fb21" (UID: "bd11183f-f4b4-4c8d-9eaf-860263e0fb21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.172415 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5xfn\" (UniqueName: \"kubernetes.io/projected/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-kube-api-access-x5xfn\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.172467 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.172485 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd11183f-f4b4-4c8d-9eaf-860263e0fb21-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.692583 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" event={"ID":"bd11183f-f4b4-4c8d-9eaf-860263e0fb21","Type":"ContainerDied","Data":"5d9f0d39bcab8ea4a3166aa17924fa22d96d9f9598a2643e953e63d1903bb862"} Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.692672 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-qtkqs" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.692709 5018 scope.go:117] "RemoveContainer" containerID="72a2313ccdc343aa69947575857f12b150ba17b3bead5ec4734ffa7051a61ad8" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.694992 5018 generic.go:334] "Generic (PLEG): container finished" podID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerID="fb861d8c016bc76027416920bf928d0a18a84db9ee084f7cac20073cef09ce36" exitCode=0 Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.695049 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerDied","Data":"fb861d8c016bc76027416920bf928d0a18a84db9ee084f7cac20073cef09ce36"} Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.716230 5018 scope.go:117] "RemoveContainer" containerID="f0b13b23065addf5b99e5a14ff66154b7d69fff3f698ea5629035c29186c1000" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.768129 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.856225 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.863461 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-qtkqs"] Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883160 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6kxd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883214 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883403 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883454 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883499 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883525 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883552 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883649 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883671 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883726 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.883773 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd\") pod \"5fd9eca2-b576-4f12-a91e-df9042db3c44\" (UID: \"5fd9eca2-b576-4f12-a91e-df9042db3c44\") " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.884221 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.884402 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.884494 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.895863 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989" (OuterVolumeSpecName: "persistence") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.896186 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info" (OuterVolumeSpecName: "pod-info") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.907687 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data" (OuterVolumeSpecName: "config-data") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.912291 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.912367 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.914827 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd" (OuterVolumeSpecName: "kube-api-access-q6kxd") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "kube-api-access-q6kxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.934849 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf" (OuterVolumeSpecName: "server-conf") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.969089 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5fd9eca2-b576-4f12-a91e-df9042db3c44" (UID: "5fd9eca2-b576-4f12-a91e-df9042db3c44"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985720 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5fd9eca2-b576-4f12-a91e-df9042db3c44-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985791 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985801 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985812 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985822 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6kxd\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-kube-api-access-q6kxd\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985833 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985865 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") on node \"crc\" " Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985876 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5fd9eca2-b576-4f12-a91e-df9042db3c44-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985885 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985893 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5fd9eca2-b576-4f12-a91e-df9042db3c44-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:49 crc kubenswrapper[5018]: I1014 08:31:49.985902 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5fd9eca2-b576-4f12-a91e-df9042db3c44-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.001833 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.001961 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989") on node "crc" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.088902 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.241100 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.398785 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399011 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399048 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399076 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rbgh\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399126 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399195 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399541 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399581 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399648 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399688 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399728 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie\") pod \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\" (UID: \"090c4f30-6438-4cd0-ad4d-8ab6531cfce5\") " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.399933 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.400042 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.400347 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.402844 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.403045 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info" (OuterVolumeSpecName: "pod-info") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.403301 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh" (OuterVolumeSpecName: "kube-api-access-9rbgh") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "kube-api-access-9rbgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.406397 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.412471 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461" (OuterVolumeSpecName: "persistence") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.420422 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data" (OuterVolumeSpecName: "config-data") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.446330 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf" (OuterVolumeSpecName: "server-conf") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.480861 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "090c4f30-6438-4cd0-ad4d-8ab6531cfce5" (UID: "090c4f30-6438-4cd0-ad4d-8ab6531cfce5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501451 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501488 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501496 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501504 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501514 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501523 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501532 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501542 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501574 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") on node \"crc\" " Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501584 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.501593 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rbgh\" (UniqueName: \"kubernetes.io/projected/090c4f30-6438-4cd0-ad4d-8ab6531cfce5-kube-api-access-9rbgh\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.519107 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.519299 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461") on node "crc" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.602122 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.619455 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" path="/var/lib/kubelet/pods/bd11183f-f4b4-4c8d-9eaf-860263e0fb21/volumes" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.707843 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.707863 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5fd9eca2-b576-4f12-a91e-df9042db3c44","Type":"ContainerDied","Data":"5f2c0a7c8618d3843e388d3fbe79a3bdfccd99fa6d1eb715215d662a368fd797"} Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.707936 5018 scope.go:117] "RemoveContainer" containerID="fb861d8c016bc76027416920bf928d0a18a84db9ee084f7cac20073cef09ce36" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.715569 5018 generic.go:334] "Generic (PLEG): container finished" podID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerID="f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b" exitCode=0 Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.715686 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.715725 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerDied","Data":"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b"} Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.716274 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090c4f30-6438-4cd0-ad4d-8ab6531cfce5","Type":"ContainerDied","Data":"7925c82d00af014ecaac9bc8d3bd931652f5b350ecf6089477413971ba9b14fa"} Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.751964 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.756434 5018 scope.go:117] "RemoveContainer" containerID="652b7b782181f73ef165e88fbce3123b71544d2476428f611fc5be1e48289969" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.770010 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.796009 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.813794 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.814003 5018 scope.go:117] "RemoveContainer" containerID="f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.823385 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.823698 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.823715 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.823727 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="init" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.823750 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="init" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.823764 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="setup-container" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.823771 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="setup-container" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.823793 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="dnsmasq-dns" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.823799 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="dnsmasq-dns" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.823812 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.824306 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.824328 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="setup-container" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.824336 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="setup-container" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.824474 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.824490 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" containerName="rabbitmq" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.824497 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd11183f-f4b4-4c8d-9eaf-860263e0fb21" containerName="dnsmasq-dns" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.825187 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.825268 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.836994 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.837260 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8c9vp" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.839113 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.839851 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.841123 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.842446 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.842489 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.845882 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.846448 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.851890 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.851929 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.852096 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.852297 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.852376 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4qpkf" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.852431 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.852585 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.855493 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.870220 5018 scope.go:117] "RemoveContainer" containerID="a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.891877 5018 scope.go:117] "RemoveContainer" containerID="f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.893100 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b\": container with ID starting with f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b not found: ID does not exist" containerID="f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.893159 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b"} err="failed to get container status \"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b\": rpc error: code = NotFound desc = could not find container \"f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b\": container with ID starting with f57e0aa635d21b8faa7e54d7def9bd0f6fd15ade4a5a971bf39edf84c5e9f18b not found: ID does not exist" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.893198 5018 scope.go:117] "RemoveContainer" containerID="a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a" Oct 14 08:31:50 crc kubenswrapper[5018]: E1014 08:31:50.893657 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a\": container with ID starting with a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a not found: ID does not exist" containerID="a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a" Oct 14 08:31:50 crc kubenswrapper[5018]: I1014 08:31:50.893784 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a"} err="failed to get container status \"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a\": rpc error: code = NotFound desc = could not find container \"a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a\": container with ID starting with a25df849b8aa9ccde544aa3d65595894432bdfe5c0a0802ff71b84e37181454a not found: ID does not exist" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007145 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007234 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007287 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007368 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007451 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007514 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007564 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007684 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kr8g\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007804 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007865 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007906 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4kbv\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.007995 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008025 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008095 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008202 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008324 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008538 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008667 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008763 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008803 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.008935 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.009036 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.110859 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.110986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111032 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4kbv\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111066 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111098 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111136 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111173 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.111210 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112131 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112182 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112286 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112387 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112435 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112465 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112547 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112696 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112775 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112834 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112897 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.112970 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113051 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113114 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113174 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kr8g\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113316 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113373 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.113175 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.114007 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.114033 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/415a1d7451e47b1410263ebc620512fbbd328bf639d64d5b904d6662fd2628ee/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.114258 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.114351 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.114871 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.115238 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.116220 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.117860 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.120151 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.121200 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.121485 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.124173 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.124231 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/927488b0368b94c1e5e36dc99b7c109dfb844e110a1f467b4cc605c2894a8706/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.124167 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.125005 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.130501 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.133275 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.136458 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4kbv\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.152287 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kr8g\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.156817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"rabbitmq-server-0\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.166149 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"rabbitmq-cell1-server-0\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.178659 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.213591 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.431496 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:31:51 crc kubenswrapper[5018]: W1014 08:31:51.443880 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29481fb2_99c9_41f0_b650_a971fa5ff28b.slice/crio-546f19dde3cba5be5adacabfec549dd44cba41c39fcbebee003e9c99b62e53b9 WatchSource:0}: Error finding container 546f19dde3cba5be5adacabfec549dd44cba41c39fcbebee003e9c99b62e53b9: Status 404 returned error can't find the container with id 546f19dde3cba5be5adacabfec549dd44cba41c39fcbebee003e9c99b62e53b9 Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.481094 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:31:51 crc kubenswrapper[5018]: W1014 08:31:51.483541 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44c35a1d_a43f_4503_a378_4149e63f8e25.slice/crio-6ac8491691b7c1d72c212c80aa65eb7ffd3f143eb7c09d519fbe409cc726bef0 WatchSource:0}: Error finding container 6ac8491691b7c1d72c212c80aa65eb7ffd3f143eb7c09d519fbe409cc726bef0: Status 404 returned error can't find the container with id 6ac8491691b7c1d72c212c80aa65eb7ffd3f143eb7c09d519fbe409cc726bef0 Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.737226 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerStarted","Data":"6ac8491691b7c1d72c212c80aa65eb7ffd3f143eb7c09d519fbe409cc726bef0"} Oct 14 08:31:51 crc kubenswrapper[5018]: I1014 08:31:51.742223 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerStarted","Data":"546f19dde3cba5be5adacabfec549dd44cba41c39fcbebee003e9c99b62e53b9"} Oct 14 08:31:52 crc kubenswrapper[5018]: I1014 08:31:52.616723 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="090c4f30-6438-4cd0-ad4d-8ab6531cfce5" path="/var/lib/kubelet/pods/090c4f30-6438-4cd0-ad4d-8ab6531cfce5/volumes" Oct 14 08:31:52 crc kubenswrapper[5018]: I1014 08:31:52.678390 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd9eca2-b576-4f12-a91e-df9042db3c44" path="/var/lib/kubelet/pods/5fd9eca2-b576-4f12-a91e-df9042db3c44/volumes" Oct 14 08:31:53 crc kubenswrapper[5018]: I1014 08:31:53.772799 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerStarted","Data":"abe3f15398c9cdd459afab86842e412ed3e64bbadd7f82be725fd3caaad3b66c"} Oct 14 08:31:53 crc kubenswrapper[5018]: I1014 08:31:53.776007 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerStarted","Data":"f4d99450585994209d83212376bd59bf0d8a61b2535a8ac45829db06d6c64d20"} Oct 14 08:32:02 crc kubenswrapper[5018]: I1014 08:32:02.463529 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:32:02 crc kubenswrapper[5018]: I1014 08:32:02.464260 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:32:27 crc kubenswrapper[5018]: I1014 08:32:27.106407 5018 generic.go:334] "Generic (PLEG): container finished" podID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerID="f4d99450585994209d83212376bd59bf0d8a61b2535a8ac45829db06d6c64d20" exitCode=0 Oct 14 08:32:27 crc kubenswrapper[5018]: I1014 08:32:27.106523 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerDied","Data":"f4d99450585994209d83212376bd59bf0d8a61b2535a8ac45829db06d6c64d20"} Oct 14 08:32:27 crc kubenswrapper[5018]: I1014 08:32:27.112718 5018 generic.go:334] "Generic (PLEG): container finished" podID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerID="abe3f15398c9cdd459afab86842e412ed3e64bbadd7f82be725fd3caaad3b66c" exitCode=0 Oct 14 08:32:27 crc kubenswrapper[5018]: I1014 08:32:27.112768 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerDied","Data":"abe3f15398c9cdd459afab86842e412ed3e64bbadd7f82be725fd3caaad3b66c"} Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.123749 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerStarted","Data":"353c1ff5f1490343dd0dfdf4922745f628e70fdbb923936bb2cb5ed954c5db54"} Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.124831 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.126663 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerStarted","Data":"efb355b1080edb66f752f3f04a2f8a2c9ade67592342b2f5278487f5bddf2b1a"} Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.126927 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.172162 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.172137866 podStartE2EDuration="38.172137866s" podCreationTimestamp="2025-10-14 08:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:32:28.164883788 +0000 UTC m=+6164.748930465" watchObservedRunningTime="2025-10-14 08:32:28.172137866 +0000 UTC m=+6164.756184523" Oct 14 08:32:28 crc kubenswrapper[5018]: I1014 08:32:28.192744 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.192713787 podStartE2EDuration="38.192713787s" podCreationTimestamp="2025-10-14 08:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:32:28.188895348 +0000 UTC m=+6164.772942025" watchObservedRunningTime="2025-10-14 08:32:28.192713787 +0000 UTC m=+6164.776760464" Oct 14 08:32:32 crc kubenswrapper[5018]: I1014 08:32:32.463025 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:32:32 crc kubenswrapper[5018]: I1014 08:32:32.463817 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:32:41 crc kubenswrapper[5018]: I1014 08:32:41.184053 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:32:41 crc kubenswrapper[5018]: I1014 08:32:41.218086 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.003657 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.006447 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.010339 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-k6crk" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.013492 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.111725 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr6bw\" (UniqueName: \"kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw\") pod \"mariadb-client-1-default\" (UID: \"22f035c1-9ee6-45ab-97ab-b46fdd58270a\") " pod="openstack/mariadb-client-1-default" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.213492 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr6bw\" (UniqueName: \"kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw\") pod \"mariadb-client-1-default\" (UID: \"22f035c1-9ee6-45ab-97ab-b46fdd58270a\") " pod="openstack/mariadb-client-1-default" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.237710 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr6bw\" (UniqueName: \"kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw\") pod \"mariadb-client-1-default\" (UID: \"22f035c1-9ee6-45ab-97ab-b46fdd58270a\") " pod="openstack/mariadb-client-1-default" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.344877 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:32:46 crc kubenswrapper[5018]: I1014 08:32:46.962903 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:32:47 crc kubenswrapper[5018]: I1014 08:32:47.315565 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"22f035c1-9ee6-45ab-97ab-b46fdd58270a","Type":"ContainerStarted","Data":"e789009b819ab4212cfbd164c665807494ab378cfc50c7fea36a56d1f8f749a6"} Oct 14 08:32:48 crc kubenswrapper[5018]: I1014 08:32:48.329590 5018 generic.go:334] "Generic (PLEG): container finished" podID="22f035c1-9ee6-45ab-97ab-b46fdd58270a" containerID="69ff4ba3ad66cda13718843584a3df7cf17247c266293e8d8583732dc0ca161e" exitCode=0 Oct 14 08:32:48 crc kubenswrapper[5018]: I1014 08:32:48.329670 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"22f035c1-9ee6-45ab-97ab-b46fdd58270a","Type":"ContainerDied","Data":"69ff4ba3ad66cda13718843584a3df7cf17247c266293e8d8583732dc0ca161e"} Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.818267 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.878704 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr6bw\" (UniqueName: \"kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw\") pod \"22f035c1-9ee6-45ab-97ab-b46fdd58270a\" (UID: \"22f035c1-9ee6-45ab-97ab-b46fdd58270a\") " Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.883908 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_22f035c1-9ee6-45ab-97ab-b46fdd58270a/mariadb-client-1-default/0.log" Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.889076 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw" (OuterVolumeSpecName: "kube-api-access-hr6bw") pod "22f035c1-9ee6-45ab-97ab-b46fdd58270a" (UID: "22f035c1-9ee6-45ab-97ab-b46fdd58270a"). InnerVolumeSpecName "kube-api-access-hr6bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.919007 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.925132 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:32:49 crc kubenswrapper[5018]: I1014 08:32:49.980682 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr6bw\" (UniqueName: \"kubernetes.io/projected/22f035c1-9ee6-45ab-97ab-b46fdd58270a-kube-api-access-hr6bw\") on node \"crc\" DevicePath \"\"" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.350015 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e789009b819ab4212cfbd164c665807494ab378cfc50c7fea36a56d1f8f749a6" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.350493 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.471705 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:32:50 crc kubenswrapper[5018]: E1014 08:32:50.472211 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22f035c1-9ee6-45ab-97ab-b46fdd58270a" containerName="mariadb-client-1-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.472242 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="22f035c1-9ee6-45ab-97ab-b46fdd58270a" containerName="mariadb-client-1-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.472507 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="22f035c1-9ee6-45ab-97ab-b46fdd58270a" containerName="mariadb-client-1-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.473461 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.476941 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-k6crk" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.484310 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.590139 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsrzz\" (UniqueName: \"kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz\") pod \"mariadb-client-2-default\" (UID: \"2ef5d0e3-2a76-435f-95bf-880361cdffe6\") " pod="openstack/mariadb-client-2-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.617713 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22f035c1-9ee6-45ab-97ab-b46fdd58270a" path="/var/lib/kubelet/pods/22f035c1-9ee6-45ab-97ab-b46fdd58270a/volumes" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.691557 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsrzz\" (UniqueName: \"kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz\") pod \"mariadb-client-2-default\" (UID: \"2ef5d0e3-2a76-435f-95bf-880361cdffe6\") " pod="openstack/mariadb-client-2-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.726818 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsrzz\" (UniqueName: \"kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz\") pod \"mariadb-client-2-default\" (UID: \"2ef5d0e3-2a76-435f-95bf-880361cdffe6\") " pod="openstack/mariadb-client-2-default" Oct 14 08:32:50 crc kubenswrapper[5018]: I1014 08:32:50.792095 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:32:51 crc kubenswrapper[5018]: I1014 08:32:51.390898 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:32:52 crc kubenswrapper[5018]: I1014 08:32:52.376860 5018 generic.go:334] "Generic (PLEG): container finished" podID="2ef5d0e3-2a76-435f-95bf-880361cdffe6" containerID="6e923d496d75ca508f131274a4ffd9d24469d09bdaa1607383b92de1e3c619f7" exitCode=0 Oct 14 08:32:52 crc kubenswrapper[5018]: I1014 08:32:52.377003 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"2ef5d0e3-2a76-435f-95bf-880361cdffe6","Type":"ContainerDied","Data":"6e923d496d75ca508f131274a4ffd9d24469d09bdaa1607383b92de1e3c619f7"} Oct 14 08:32:52 crc kubenswrapper[5018]: I1014 08:32:52.377344 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"2ef5d0e3-2a76-435f-95bf-880361cdffe6","Type":"ContainerStarted","Data":"67870c7ce4dafa0ec5a62cfeae471a9b2a570ffe93acd849868ff25422adabcc"} Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.775140 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.829288 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_2ef5d0e3-2a76-435f-95bf-880361cdffe6/mariadb-client-2-default/0.log" Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.845708 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsrzz\" (UniqueName: \"kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz\") pod \"2ef5d0e3-2a76-435f-95bf-880361cdffe6\" (UID: \"2ef5d0e3-2a76-435f-95bf-880361cdffe6\") " Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.855517 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz" (OuterVolumeSpecName: "kube-api-access-qsrzz") pod "2ef5d0e3-2a76-435f-95bf-880361cdffe6" (UID: "2ef5d0e3-2a76-435f-95bf-880361cdffe6"). InnerVolumeSpecName "kube-api-access-qsrzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.875425 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.886568 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:32:53 crc kubenswrapper[5018]: I1014 08:32:53.948036 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsrzz\" (UniqueName: \"kubernetes.io/projected/2ef5d0e3-2a76-435f-95bf-880361cdffe6-kube-api-access-qsrzz\") on node \"crc\" DevicePath \"\"" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.025399 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-3-default"] Oct 14 08:32:54 crc kubenswrapper[5018]: E1014 08:32:54.025986 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef5d0e3-2a76-435f-95bf-880361cdffe6" containerName="mariadb-client-2-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.026006 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef5d0e3-2a76-435f-95bf-880361cdffe6" containerName="mariadb-client-2-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.026232 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef5d0e3-2a76-435f-95bf-880361cdffe6" containerName="mariadb-client-2-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.026797 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.038035 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.050227 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j9mt\" (UniqueName: \"kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt\") pod \"mariadb-client-3-default\" (UID: \"08b78800-148f-4385-8763-5b45f57253b8\") " pod="openstack/mariadb-client-3-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.152297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j9mt\" (UniqueName: \"kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt\") pod \"mariadb-client-3-default\" (UID: \"08b78800-148f-4385-8763-5b45f57253b8\") " pod="openstack/mariadb-client-3-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.174702 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j9mt\" (UniqueName: \"kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt\") pod \"mariadb-client-3-default\" (UID: \"08b78800-148f-4385-8763-5b45f57253b8\") " pod="openstack/mariadb-client-3-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.351080 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.397737 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67870c7ce4dafa0ec5a62cfeae471a9b2a570ffe93acd849868ff25422adabcc" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.397837 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.636328 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef5d0e3-2a76-435f-95bf-880361cdffe6" path="/var/lib/kubelet/pods/2ef5d0e3-2a76-435f-95bf-880361cdffe6/volumes" Oct 14 08:32:54 crc kubenswrapper[5018]: I1014 08:32:54.961370 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 14 08:32:54 crc kubenswrapper[5018]: W1014 08:32:54.961890 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08b78800_148f_4385_8763_5b45f57253b8.slice/crio-e7792b0c4c111f735ab8a1b312b0595dcfb06e211e0186ca1b49c3a1a538796e WatchSource:0}: Error finding container e7792b0c4c111f735ab8a1b312b0595dcfb06e211e0186ca1b49c3a1a538796e: Status 404 returned error can't find the container with id e7792b0c4c111f735ab8a1b312b0595dcfb06e211e0186ca1b49c3a1a538796e Oct 14 08:32:55 crc kubenswrapper[5018]: I1014 08:32:55.409549 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"08b78800-148f-4385-8763-5b45f57253b8","Type":"ContainerStarted","Data":"3403ddcf9fd9fac69da8b33a0da5faef1eee0bd42b65e205336b4f8ac5366607"} Oct 14 08:32:55 crc kubenswrapper[5018]: I1014 08:32:55.409990 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"08b78800-148f-4385-8763-5b45f57253b8","Type":"ContainerStarted","Data":"e7792b0c4c111f735ab8a1b312b0595dcfb06e211e0186ca1b49c3a1a538796e"} Oct 14 08:32:55 crc kubenswrapper[5018]: I1014 08:32:55.439768 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-3-default" podStartSLOduration=1.439741055 podStartE2EDuration="1.439741055s" podCreationTimestamp="2025-10-14 08:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:32:55.427148833 +0000 UTC m=+6192.011195510" watchObservedRunningTime="2025-10-14 08:32:55.439741055 +0000 UTC m=+6192.023787722" Oct 14 08:32:57 crc kubenswrapper[5018]: I1014 08:32:57.437497 5018 generic.go:334] "Generic (PLEG): container finished" podID="08b78800-148f-4385-8763-5b45f57253b8" containerID="3403ddcf9fd9fac69da8b33a0da5faef1eee0bd42b65e205336b4f8ac5366607" exitCode=0 Oct 14 08:32:57 crc kubenswrapper[5018]: I1014 08:32:57.437720 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"08b78800-148f-4385-8763-5b45f57253b8","Type":"ContainerDied","Data":"3403ddcf9fd9fac69da8b33a0da5faef1eee0bd42b65e205336b4f8ac5366607"} Oct 14 08:32:58 crc kubenswrapper[5018]: I1014 08:32:58.861206 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 14 08:32:58 crc kubenswrapper[5018]: I1014 08:32:58.925754 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 14 08:32:58 crc kubenswrapper[5018]: I1014 08:32:58.930914 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.035003 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j9mt\" (UniqueName: \"kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt\") pod \"08b78800-148f-4385-8763-5b45f57253b8\" (UID: \"08b78800-148f-4385-8763-5b45f57253b8\") " Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.043322 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt" (OuterVolumeSpecName: "kube-api-access-6j9mt") pod "08b78800-148f-4385-8763-5b45f57253b8" (UID: "08b78800-148f-4385-8763-5b45f57253b8"). InnerVolumeSpecName "kube-api-access-6j9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.143946 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j9mt\" (UniqueName: \"kubernetes.io/projected/08b78800-148f-4385-8763-5b45f57253b8-kube-api-access-6j9mt\") on node \"crc\" DevicePath \"\"" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.444102 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:32:59 crc kubenswrapper[5018]: E1014 08:32:59.445244 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b78800-148f-4385-8763-5b45f57253b8" containerName="mariadb-client-3-default" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.445282 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b78800-148f-4385-8763-5b45f57253b8" containerName="mariadb-client-3-default" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.445573 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b78800-148f-4385-8763-5b45f57253b8" containerName="mariadb-client-3-default" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.446527 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.455245 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2tn7\" (UniqueName: \"kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7\") pod \"mariadb-client-1\" (UID: \"6b9b2dfd-c297-417e-ab85-80ef4685a4c7\") " pod="openstack/mariadb-client-1" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.459547 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.473462 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7792b0c4c111f735ab8a1b312b0595dcfb06e211e0186ca1b49c3a1a538796e" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.473571 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.556768 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2tn7\" (UniqueName: \"kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7\") pod \"mariadb-client-1\" (UID: \"6b9b2dfd-c297-417e-ab85-80ef4685a4c7\") " pod="openstack/mariadb-client-1" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.580301 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2tn7\" (UniqueName: \"kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7\") pod \"mariadb-client-1\" (UID: \"6b9b2dfd-c297-417e-ab85-80ef4685a4c7\") " pod="openstack/mariadb-client-1" Oct 14 08:32:59 crc kubenswrapper[5018]: I1014 08:32:59.783940 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:33:00 crc kubenswrapper[5018]: I1014 08:33:00.177192 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:33:00 crc kubenswrapper[5018]: W1014 08:33:00.190839 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b9b2dfd_c297_417e_ab85_80ef4685a4c7.slice/crio-78bd9a2e1586ecebd844c79ab8bfc67eb6f9dda0bcf3d8074dcb4fc73edab4ce WatchSource:0}: Error finding container 78bd9a2e1586ecebd844c79ab8bfc67eb6f9dda0bcf3d8074dcb4fc73edab4ce: Status 404 returned error can't find the container with id 78bd9a2e1586ecebd844c79ab8bfc67eb6f9dda0bcf3d8074dcb4fc73edab4ce Oct 14 08:33:00 crc kubenswrapper[5018]: I1014 08:33:00.483470 5018 generic.go:334] "Generic (PLEG): container finished" podID="6b9b2dfd-c297-417e-ab85-80ef4685a4c7" containerID="9b302e8f1cfd864931de5f6f4bec13f42394bd3439340f88cd73d1a0726ab015" exitCode=0 Oct 14 08:33:00 crc kubenswrapper[5018]: I1014 08:33:00.483532 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"6b9b2dfd-c297-417e-ab85-80ef4685a4c7","Type":"ContainerDied","Data":"9b302e8f1cfd864931de5f6f4bec13f42394bd3439340f88cd73d1a0726ab015"} Oct 14 08:33:00 crc kubenswrapper[5018]: I1014 08:33:00.483575 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"6b9b2dfd-c297-417e-ab85-80ef4685a4c7","Type":"ContainerStarted","Data":"78bd9a2e1586ecebd844c79ab8bfc67eb6f9dda0bcf3d8074dcb4fc73edab4ce"} Oct 14 08:33:00 crc kubenswrapper[5018]: I1014 08:33:00.620591 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b78800-148f-4385-8763-5b45f57253b8" path="/var/lib/kubelet/pods/08b78800-148f-4385-8763-5b45f57253b8/volumes" Oct 14 08:33:01 crc kubenswrapper[5018]: I1014 08:33:01.963365 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:33:01 crc kubenswrapper[5018]: I1014 08:33:01.986370 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_6b9b2dfd-c297-417e-ab85-80ef4685a4c7/mariadb-client-1/0.log" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.022526 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.029450 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.100587 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2tn7\" (UniqueName: \"kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7\") pod \"6b9b2dfd-c297-417e-ab85-80ef4685a4c7\" (UID: \"6b9b2dfd-c297-417e-ab85-80ef4685a4c7\") " Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.109819 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7" (OuterVolumeSpecName: "kube-api-access-p2tn7") pod "6b9b2dfd-c297-417e-ab85-80ef4685a4c7" (UID: "6b9b2dfd-c297-417e-ab85-80ef4685a4c7"). InnerVolumeSpecName "kube-api-access-p2tn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.202881 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2tn7\" (UniqueName: \"kubernetes.io/projected/6b9b2dfd-c297-417e-ab85-80ef4685a4c7-kube-api-access-p2tn7\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.463504 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.463855 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.464097 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.465102 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.465333 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99" gracePeriod=600 Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.498443 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:33:02 crc kubenswrapper[5018]: E1014 08:33:02.498848 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9b2dfd-c297-417e-ab85-80ef4685a4c7" containerName="mariadb-client-1" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.498879 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9b2dfd-c297-417e-ab85-80ef4685a4c7" containerName="mariadb-client-1" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.499064 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9b2dfd-c297-417e-ab85-80ef4685a4c7" containerName="mariadb-client-1" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.499725 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.502099 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78bd9a2e1586ecebd844c79ab8bfc67eb6f9dda0bcf3d8074dcb4fc73edab4ce" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.502164 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.506827 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9sj8\" (UniqueName: \"kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8\") pod \"mariadb-client-4-default\" (UID: \"c0123171-616d-44f1-b95a-a1c6e7d29e59\") " pod="openstack/mariadb-client-4-default" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.509737 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.607816 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9sj8\" (UniqueName: \"kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8\") pod \"mariadb-client-4-default\" (UID: \"c0123171-616d-44f1-b95a-a1c6e7d29e59\") " pod="openstack/mariadb-client-4-default" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.623593 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b9b2dfd-c297-417e-ab85-80ef4685a4c7" path="/var/lib/kubelet/pods/6b9b2dfd-c297-417e-ab85-80ef4685a4c7/volumes" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.630886 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9sj8\" (UniqueName: \"kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8\") pod \"mariadb-client-4-default\" (UID: \"c0123171-616d-44f1-b95a-a1c6e7d29e59\") " pod="openstack/mariadb-client-4-default" Oct 14 08:33:02 crc kubenswrapper[5018]: I1014 08:33:02.866478 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.228804 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.515226 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99" exitCode=0 Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.515291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99"} Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.515315 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c"} Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.515330 5018 scope.go:117] "RemoveContainer" containerID="5d278380f873454f9cef63bb02b874475b1ff5b913d2c20f177d9d9cdc289d78" Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.519286 5018 generic.go:334] "Generic (PLEG): container finished" podID="c0123171-616d-44f1-b95a-a1c6e7d29e59" containerID="2a0ba0e528c7064576ebd5cf88b5521b454b3b9271d987d18ec702cc760b8c61" exitCode=0 Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.519326 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"c0123171-616d-44f1-b95a-a1c6e7d29e59","Type":"ContainerDied","Data":"2a0ba0e528c7064576ebd5cf88b5521b454b3b9271d987d18ec702cc760b8c61"} Oct 14 08:33:03 crc kubenswrapper[5018]: I1014 08:33:03.519351 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"c0123171-616d-44f1-b95a-a1c6e7d29e59","Type":"ContainerStarted","Data":"3ce592a7d9142c9859dbd18cbb5e4f7d9a89b70dc557b5e083262ef241db1e28"} Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.080947 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.113107 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_c0123171-616d-44f1-b95a-a1c6e7d29e59/mariadb-client-4-default/0.log" Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.138176 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.143088 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.268398 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9sj8\" (UniqueName: \"kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8\") pod \"c0123171-616d-44f1-b95a-a1c6e7d29e59\" (UID: \"c0123171-616d-44f1-b95a-a1c6e7d29e59\") " Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.275859 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8" (OuterVolumeSpecName: "kube-api-access-s9sj8") pod "c0123171-616d-44f1-b95a-a1c6e7d29e59" (UID: "c0123171-616d-44f1-b95a-a1c6e7d29e59"). InnerVolumeSpecName "kube-api-access-s9sj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.370884 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9sj8\" (UniqueName: \"kubernetes.io/projected/c0123171-616d-44f1-b95a-a1c6e7d29e59-kube-api-access-s9sj8\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.561251 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ce592a7d9142c9859dbd18cbb5e4f7d9a89b70dc557b5e083262ef241db1e28" Oct 14 08:33:05 crc kubenswrapper[5018]: I1014 08:33:05.561681 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:33:06 crc kubenswrapper[5018]: I1014 08:33:06.616757 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0123171-616d-44f1-b95a-a1c6e7d29e59" path="/var/lib/kubelet/pods/c0123171-616d-44f1-b95a-a1c6e7d29e59/volumes" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.033847 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:33:09 crc kubenswrapper[5018]: E1014 08:33:09.034899 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0123171-616d-44f1-b95a-a1c6e7d29e59" containerName="mariadb-client-4-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.034918 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0123171-616d-44f1-b95a-a1c6e7d29e59" containerName="mariadb-client-4-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.035141 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0123171-616d-44f1-b95a-a1c6e7d29e59" containerName="mariadb-client-4-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.035869 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.038589 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-k6crk" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.045484 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.136115 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kvgl\" (UniqueName: \"kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl\") pod \"mariadb-client-5-default\" (UID: \"afa10410-43f5-4470-a2fc-ac0c3a02e695\") " pod="openstack/mariadb-client-5-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.238353 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kvgl\" (UniqueName: \"kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl\") pod \"mariadb-client-5-default\" (UID: \"afa10410-43f5-4470-a2fc-ac0c3a02e695\") " pod="openstack/mariadb-client-5-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.259132 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kvgl\" (UniqueName: \"kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl\") pod \"mariadb-client-5-default\" (UID: \"afa10410-43f5-4470-a2fc-ac0c3a02e695\") " pod="openstack/mariadb-client-5-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.375894 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:33:09 crc kubenswrapper[5018]: I1014 08:33:09.715325 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:33:10 crc kubenswrapper[5018]: I1014 08:33:10.605232 5018 generic.go:334] "Generic (PLEG): container finished" podID="afa10410-43f5-4470-a2fc-ac0c3a02e695" containerID="7ed138fc940fc3e7de7d8c749f8425217059cb4f99a01e799475f6a12e07a897" exitCode=0 Oct 14 08:33:10 crc kubenswrapper[5018]: I1014 08:33:10.615450 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"afa10410-43f5-4470-a2fc-ac0c3a02e695","Type":"ContainerDied","Data":"7ed138fc940fc3e7de7d8c749f8425217059cb4f99a01e799475f6a12e07a897"} Oct 14 08:33:10 crc kubenswrapper[5018]: I1014 08:33:10.615486 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"afa10410-43f5-4470-a2fc-ac0c3a02e695","Type":"ContainerStarted","Data":"a57c78888c7e2676f2bf263d83673e9477e6ac8d034c6658a9775e910ed31a25"} Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.061048 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.081531 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_afa10410-43f5-4470-a2fc-ac0c3a02e695/mariadb-client-5-default/0.log" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.089817 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kvgl\" (UniqueName: \"kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl\") pod \"afa10410-43f5-4470-a2fc-ac0c3a02e695\" (UID: \"afa10410-43f5-4470-a2fc-ac0c3a02e695\") " Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.096288 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl" (OuterVolumeSpecName: "kube-api-access-5kvgl") pod "afa10410-43f5-4470-a2fc-ac0c3a02e695" (UID: "afa10410-43f5-4470-a2fc-ac0c3a02e695"). InnerVolumeSpecName "kube-api-access-5kvgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.105574 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.110294 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.191131 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kvgl\" (UniqueName: \"kubernetes.io/projected/afa10410-43f5-4470-a2fc-ac0c3a02e695-kube-api-access-5kvgl\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.291423 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:33:12 crc kubenswrapper[5018]: E1014 08:33:12.291819 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa10410-43f5-4470-a2fc-ac0c3a02e695" containerName="mariadb-client-5-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.291835 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa10410-43f5-4470-a2fc-ac0c3a02e695" containerName="mariadb-client-5-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.292018 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa10410-43f5-4470-a2fc-ac0c3a02e695" containerName="mariadb-client-5-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.292560 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.296451 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.394287 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd6fh\" (UniqueName: \"kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh\") pod \"mariadb-client-6-default\" (UID: \"e98d6afa-6e6a-4a59-9443-90e90a4195bc\") " pod="openstack/mariadb-client-6-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.495996 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd6fh\" (UniqueName: \"kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh\") pod \"mariadb-client-6-default\" (UID: \"e98d6afa-6e6a-4a59-9443-90e90a4195bc\") " pod="openstack/mariadb-client-6-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.530477 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd6fh\" (UniqueName: \"kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh\") pod \"mariadb-client-6-default\" (UID: \"e98d6afa-6e6a-4a59-9443-90e90a4195bc\") " pod="openstack/mariadb-client-6-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.615990 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.628690 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa10410-43f5-4470-a2fc-ac0c3a02e695" path="/var/lib/kubelet/pods/afa10410-43f5-4470-a2fc-ac0c3a02e695/volumes" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.652964 5018 scope.go:117] "RemoveContainer" containerID="7ed138fc940fc3e7de7d8c749f8425217059cb4f99a01e799475f6a12e07a897" Oct 14 08:33:12 crc kubenswrapper[5018]: I1014 08:33:12.653001 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:33:13 crc kubenswrapper[5018]: I1014 08:33:13.240803 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:33:13 crc kubenswrapper[5018]: W1014 08:33:13.254916 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode98d6afa_6e6a_4a59_9443_90e90a4195bc.slice/crio-2bf968a0a500fd5da33afa8458532acc26eca3a5d4f370cd5dbd0c3b03c6d9f4 WatchSource:0}: Error finding container 2bf968a0a500fd5da33afa8458532acc26eca3a5d4f370cd5dbd0c3b03c6d9f4: Status 404 returned error can't find the container with id 2bf968a0a500fd5da33afa8458532acc26eca3a5d4f370cd5dbd0c3b03c6d9f4 Oct 14 08:33:13 crc kubenswrapper[5018]: I1014 08:33:13.660557 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"e98d6afa-6e6a-4a59-9443-90e90a4195bc","Type":"ContainerStarted","Data":"8cb810321b69caa6723c48f0b5e918ba68976e43d9eaffbd97f58fe97744f9dd"} Oct 14 08:33:13 crc kubenswrapper[5018]: I1014 08:33:13.660890 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"e98d6afa-6e6a-4a59-9443-90e90a4195bc","Type":"ContainerStarted","Data":"2bf968a0a500fd5da33afa8458532acc26eca3a5d4f370cd5dbd0c3b03c6d9f4"} Oct 14 08:33:13 crc kubenswrapper[5018]: I1014 08:33:13.675281 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.675262316 podStartE2EDuration="1.675262316s" podCreationTimestamp="2025-10-14 08:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:33:13.674089232 +0000 UTC m=+6210.258135849" watchObservedRunningTime="2025-10-14 08:33:13.675262316 +0000 UTC m=+6210.259308943" Oct 14 08:33:14 crc kubenswrapper[5018]: I1014 08:33:14.674300 5018 generic.go:334] "Generic (PLEG): container finished" podID="e98d6afa-6e6a-4a59-9443-90e90a4195bc" containerID="8cb810321b69caa6723c48f0b5e918ba68976e43d9eaffbd97f58fe97744f9dd" exitCode=0 Oct 14 08:33:14 crc kubenswrapper[5018]: I1014 08:33:14.674344 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"e98d6afa-6e6a-4a59-9443-90e90a4195bc","Type":"ContainerDied","Data":"8cb810321b69caa6723c48f0b5e918ba68976e43d9eaffbd97f58fe97744f9dd"} Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.096960 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.135072 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.140232 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.152172 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd6fh\" (UniqueName: \"kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh\") pod \"e98d6afa-6e6a-4a59-9443-90e90a4195bc\" (UID: \"e98d6afa-6e6a-4a59-9443-90e90a4195bc\") " Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.157326 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh" (OuterVolumeSpecName: "kube-api-access-fd6fh") pod "e98d6afa-6e6a-4a59-9443-90e90a4195bc" (UID: "e98d6afa-6e6a-4a59-9443-90e90a4195bc"). InnerVolumeSpecName "kube-api-access-fd6fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.253562 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd6fh\" (UniqueName: \"kubernetes.io/projected/e98d6afa-6e6a-4a59-9443-90e90a4195bc-kube-api-access-fd6fh\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.297103 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:33:16 crc kubenswrapper[5018]: E1014 08:33:16.298153 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98d6afa-6e6a-4a59-9443-90e90a4195bc" containerName="mariadb-client-6-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.298210 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98d6afa-6e6a-4a59-9443-90e90a4195bc" containerName="mariadb-client-6-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.298555 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98d6afa-6e6a-4a59-9443-90e90a4195bc" containerName="mariadb-client-6-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.299344 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.306194 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.355275 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttl4\" (UniqueName: \"kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4\") pod \"mariadb-client-7-default\" (UID: \"908ad440-ca17-4bb9-95a4-a6d72d3d3a48\") " pod="openstack/mariadb-client-7-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.457045 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttl4\" (UniqueName: \"kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4\") pod \"mariadb-client-7-default\" (UID: \"908ad440-ca17-4bb9-95a4-a6d72d3d3a48\") " pod="openstack/mariadb-client-7-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.478589 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttl4\" (UniqueName: \"kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4\") pod \"mariadb-client-7-default\" (UID: \"908ad440-ca17-4bb9-95a4-a6d72d3d3a48\") " pod="openstack/mariadb-client-7-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.620372 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98d6afa-6e6a-4a59-9443-90e90a4195bc" path="/var/lib/kubelet/pods/e98d6afa-6e6a-4a59-9443-90e90a4195bc/volumes" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.632520 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.703709 5018 scope.go:117] "RemoveContainer" containerID="8cb810321b69caa6723c48f0b5e918ba68976e43d9eaffbd97f58fe97744f9dd" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.703969 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:33:16 crc kubenswrapper[5018]: I1014 08:33:16.990994 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:33:16 crc kubenswrapper[5018]: W1014 08:33:16.998159 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod908ad440_ca17_4bb9_95a4_a6d72d3d3a48.slice/crio-4add8490941f57b877d7eff1b624b7f3e70cf687a563201c40b39b52b0d53814 WatchSource:0}: Error finding container 4add8490941f57b877d7eff1b624b7f3e70cf687a563201c40b39b52b0d53814: Status 404 returned error can't find the container with id 4add8490941f57b877d7eff1b624b7f3e70cf687a563201c40b39b52b0d53814 Oct 14 08:33:17 crc kubenswrapper[5018]: I1014 08:33:17.719308 5018 generic.go:334] "Generic (PLEG): container finished" podID="908ad440-ca17-4bb9-95a4-a6d72d3d3a48" containerID="0a6050e9aa1aa2cc5aa7b33655406bfbdb81e224eb7c8413c00f5745b6b373e3" exitCode=0 Oct 14 08:33:17 crc kubenswrapper[5018]: I1014 08:33:17.719579 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"908ad440-ca17-4bb9-95a4-a6d72d3d3a48","Type":"ContainerDied","Data":"0a6050e9aa1aa2cc5aa7b33655406bfbdb81e224eb7c8413c00f5745b6b373e3"} Oct 14 08:33:17 crc kubenswrapper[5018]: I1014 08:33:17.719902 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"908ad440-ca17-4bb9-95a4-a6d72d3d3a48","Type":"ContainerStarted","Data":"4add8490941f57b877d7eff1b624b7f3e70cf687a563201c40b39b52b0d53814"} Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.188138 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.209099 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_908ad440-ca17-4bb9-95a4-a6d72d3d3a48/mariadb-client-7-default/0.log" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.236740 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.241695 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.315564 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ttl4\" (UniqueName: \"kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4\") pod \"908ad440-ca17-4bb9-95a4-a6d72d3d3a48\" (UID: \"908ad440-ca17-4bb9-95a4-a6d72d3d3a48\") " Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.324492 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4" (OuterVolumeSpecName: "kube-api-access-9ttl4") pod "908ad440-ca17-4bb9-95a4-a6d72d3d3a48" (UID: "908ad440-ca17-4bb9-95a4-a6d72d3d3a48"). InnerVolumeSpecName "kube-api-access-9ttl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.417792 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ttl4\" (UniqueName: \"kubernetes.io/projected/908ad440-ca17-4bb9-95a4-a6d72d3d3a48-kube-api-access-9ttl4\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.426571 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:33:19 crc kubenswrapper[5018]: E1014 08:33:19.427093 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908ad440-ca17-4bb9-95a4-a6d72d3d3a48" containerName="mariadb-client-7-default" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.427125 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="908ad440-ca17-4bb9-95a4-a6d72d3d3a48" containerName="mariadb-client-7-default" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.427433 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="908ad440-ca17-4bb9-95a4-a6d72d3d3a48" containerName="mariadb-client-7-default" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.428253 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.437086 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.520160 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlrlh\" (UniqueName: \"kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh\") pod \"mariadb-client-2\" (UID: \"3b5fb417-9e1b-48a4-9fd9-d537f495594e\") " pod="openstack/mariadb-client-2" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.622108 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlrlh\" (UniqueName: \"kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh\") pod \"mariadb-client-2\" (UID: \"3b5fb417-9e1b-48a4-9fd9-d537f495594e\") " pod="openstack/mariadb-client-2" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.653028 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlrlh\" (UniqueName: \"kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh\") pod \"mariadb-client-2\" (UID: \"3b5fb417-9e1b-48a4-9fd9-d537f495594e\") " pod="openstack/mariadb-client-2" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.742828 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4add8490941f57b877d7eff1b624b7f3e70cf687a563201c40b39b52b0d53814" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.742924 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:33:19 crc kubenswrapper[5018]: I1014 08:33:19.753023 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:33:20 crc kubenswrapper[5018]: I1014 08:33:20.450551 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:33:20 crc kubenswrapper[5018]: I1014 08:33:20.614570 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="908ad440-ca17-4bb9-95a4-a6d72d3d3a48" path="/var/lib/kubelet/pods/908ad440-ca17-4bb9-95a4-a6d72d3d3a48/volumes" Oct 14 08:33:20 crc kubenswrapper[5018]: I1014 08:33:20.754711 5018 generic.go:334] "Generic (PLEG): container finished" podID="3b5fb417-9e1b-48a4-9fd9-d537f495594e" containerID="ca3116eef6dffd5048d1a8ad98df4ea875393a58329bca22e3e592aecdfaf1cd" exitCode=0 Oct 14 08:33:20 crc kubenswrapper[5018]: I1014 08:33:20.754774 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"3b5fb417-9e1b-48a4-9fd9-d537f495594e","Type":"ContainerDied","Data":"ca3116eef6dffd5048d1a8ad98df4ea875393a58329bca22e3e592aecdfaf1cd"} Oct 14 08:33:20 crc kubenswrapper[5018]: I1014 08:33:20.754817 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"3b5fb417-9e1b-48a4-9fd9-d537f495594e","Type":"ContainerStarted","Data":"88c6a42d92911e303b5fbb75a645698e8e98f9014021e02f7553bdd8dafb0307"} Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.145441 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.166811 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_3b5fb417-9e1b-48a4-9fd9-d537f495594e/mariadb-client-2/0.log" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.193155 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.200908 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.273101 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlrlh\" (UniqueName: \"kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh\") pod \"3b5fb417-9e1b-48a4-9fd9-d537f495594e\" (UID: \"3b5fb417-9e1b-48a4-9fd9-d537f495594e\") " Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.281158 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh" (OuterVolumeSpecName: "kube-api-access-jlrlh") pod "3b5fb417-9e1b-48a4-9fd9-d537f495594e" (UID: "3b5fb417-9e1b-48a4-9fd9-d537f495594e"). InnerVolumeSpecName "kube-api-access-jlrlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.375495 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlrlh\" (UniqueName: \"kubernetes.io/projected/3b5fb417-9e1b-48a4-9fd9-d537f495594e-kube-api-access-jlrlh\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.619679 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5fb417-9e1b-48a4-9fd9-d537f495594e" path="/var/lib/kubelet/pods/3b5fb417-9e1b-48a4-9fd9-d537f495594e/volumes" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.775909 5018 scope.go:117] "RemoveContainer" containerID="ca3116eef6dffd5048d1a8ad98df4ea875393a58329bca22e3e592aecdfaf1cd" Oct 14 08:33:22 crc kubenswrapper[5018]: I1014 08:33:22.775940 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.345508 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:33:55 crc kubenswrapper[5018]: E1014 08:33:55.349486 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5fb417-9e1b-48a4-9fd9-d537f495594e" containerName="mariadb-client-2" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.349559 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5fb417-9e1b-48a4-9fd9-d537f495594e" containerName="mariadb-client-2" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.350034 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5fb417-9e1b-48a4-9fd9-d537f495594e" containerName="mariadb-client-2" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.352457 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.360531 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.400697 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.400920 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.400993 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqlb6\" (UniqueName: \"kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.502279 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.502463 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.502559 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqlb6\" (UniqueName: \"kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.502872 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.502941 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.526044 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqlb6\" (UniqueName: \"kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6\") pod \"community-operators-xzwb4\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:55 crc kubenswrapper[5018]: I1014 08:33:55.724096 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:33:56 crc kubenswrapper[5018]: I1014 08:33:56.187797 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:33:56 crc kubenswrapper[5018]: W1014 08:33:56.197119 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8954cde6_6c36_40ab_a4fb_d80ed27e4bd8.slice/crio-e5e318a80466aa62fcec610bd7bdd8c0e22eb592f961909b6ac4f503c8225771 WatchSource:0}: Error finding container e5e318a80466aa62fcec610bd7bdd8c0e22eb592f961909b6ac4f503c8225771: Status 404 returned error can't find the container with id e5e318a80466aa62fcec610bd7bdd8c0e22eb592f961909b6ac4f503c8225771 Oct 14 08:33:57 crc kubenswrapper[5018]: I1014 08:33:57.133763 5018 generic.go:334] "Generic (PLEG): container finished" podID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerID="5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136" exitCode=0 Oct 14 08:33:57 crc kubenswrapper[5018]: I1014 08:33:57.133861 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerDied","Data":"5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136"} Oct 14 08:33:57 crc kubenswrapper[5018]: I1014 08:33:57.134301 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerStarted","Data":"e5e318a80466aa62fcec610bd7bdd8c0e22eb592f961909b6ac4f503c8225771"} Oct 14 08:33:59 crc kubenswrapper[5018]: I1014 08:33:59.163921 5018 generic.go:334] "Generic (PLEG): container finished" podID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerID="889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529" exitCode=0 Oct 14 08:33:59 crc kubenswrapper[5018]: I1014 08:33:59.164007 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerDied","Data":"889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529"} Oct 14 08:34:00 crc kubenswrapper[5018]: I1014 08:34:00.179306 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerStarted","Data":"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16"} Oct 14 08:34:00 crc kubenswrapper[5018]: I1014 08:34:00.210322 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xzwb4" podStartSLOduration=2.6089462919999997 podStartE2EDuration="5.210298056s" podCreationTimestamp="2025-10-14 08:33:55 +0000 UTC" firstStartedPulling="2025-10-14 08:33:57.137687002 +0000 UTC m=+6253.721733669" lastFinishedPulling="2025-10-14 08:33:59.739038766 +0000 UTC m=+6256.323085433" observedRunningTime="2025-10-14 08:34:00.205348884 +0000 UTC m=+6256.789395551" watchObservedRunningTime="2025-10-14 08:34:00.210298056 +0000 UTC m=+6256.794344713" Oct 14 08:34:05 crc kubenswrapper[5018]: I1014 08:34:05.725179 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:05 crc kubenswrapper[5018]: I1014 08:34:05.725973 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:05 crc kubenswrapper[5018]: I1014 08:34:05.808788 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.320839 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.368409 5018 scope.go:117] "RemoveContainer" containerID="1542282a89a23a44f384197f8c3fc8e2f4bdacbb48b918bf86765243f9e25d22" Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.393637 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.408039 5018 scope.go:117] "RemoveContainer" containerID="c9585659bb37b2198dfb207da2e61c41c96bc4dc83b470d3fc11c35c1c5cdacd" Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.455838 5018 scope.go:117] "RemoveContainer" containerID="b350ff55fd93c69942fd77ca393744433abc13f145b494dd68a72c35f873eee4" Oct 14 08:34:06 crc kubenswrapper[5018]: I1014 08:34:06.494602 5018 scope.go:117] "RemoveContainer" containerID="5cd523cba377fd487712d07c88c9965638070d41087d4cfea5ce502a3248d2d9" Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.268911 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xzwb4" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="registry-server" containerID="cri-o://f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16" gracePeriod=2 Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.805035 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.911976 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqlb6\" (UniqueName: \"kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6\") pod \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.912226 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content\") pod \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.912292 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities\") pod \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\" (UID: \"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8\") " Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.914420 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities" (OuterVolumeSpecName: "utilities") pod "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" (UID: "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:34:08 crc kubenswrapper[5018]: I1014 08:34:08.921165 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6" (OuterVolumeSpecName: "kube-api-access-nqlb6") pod "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" (UID: "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8"). InnerVolumeSpecName "kube-api-access-nqlb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.014922 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.014975 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqlb6\" (UniqueName: \"kubernetes.io/projected/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-kube-api-access-nqlb6\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.025444 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" (UID: "8954cde6-6c36-40ab-a4fb-d80ed27e4bd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.116747 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.283268 5018 generic.go:334] "Generic (PLEG): container finished" podID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerID="f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16" exitCode=0 Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.283503 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerDied","Data":"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16"} Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.283552 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xzwb4" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.283585 5018 scope.go:117] "RemoveContainer" containerID="f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.283566 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xzwb4" event={"ID":"8954cde6-6c36-40ab-a4fb-d80ed27e4bd8","Type":"ContainerDied","Data":"e5e318a80466aa62fcec610bd7bdd8c0e22eb592f961909b6ac4f503c8225771"} Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.327050 5018 scope.go:117] "RemoveContainer" containerID="889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.346728 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.361237 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xzwb4"] Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.373873 5018 scope.go:117] "RemoveContainer" containerID="5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.406831 5018 scope.go:117] "RemoveContainer" containerID="f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16" Oct 14 08:34:09 crc kubenswrapper[5018]: E1014 08:34:09.410930 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16\": container with ID starting with f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16 not found: ID does not exist" containerID="f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.411292 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16"} err="failed to get container status \"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16\": rpc error: code = NotFound desc = could not find container \"f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16\": container with ID starting with f864fe86ec4c109827a79eb481066b6dee15275e3ef4ca4bf233be1a516f0a16 not found: ID does not exist" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.411326 5018 scope.go:117] "RemoveContainer" containerID="889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529" Oct 14 08:34:09 crc kubenswrapper[5018]: E1014 08:34:09.413197 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529\": container with ID starting with 889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529 not found: ID does not exist" containerID="889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.414463 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529"} err="failed to get container status \"889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529\": rpc error: code = NotFound desc = could not find container \"889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529\": container with ID starting with 889bff4288055c07073f7eae70cea5dc4454aea12364a22676dc8877520a6529 not found: ID does not exist" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.414539 5018 scope.go:117] "RemoveContainer" containerID="5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136" Oct 14 08:34:09 crc kubenswrapper[5018]: E1014 08:34:09.414965 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136\": container with ID starting with 5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136 not found: ID does not exist" containerID="5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136" Oct 14 08:34:09 crc kubenswrapper[5018]: I1014 08:34:09.415014 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136"} err="failed to get container status \"5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136\": rpc error: code = NotFound desc = could not find container \"5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136\": container with ID starting with 5ea4d6f81c96f6661ae615f37245acf653392f53be31b1db698f9e130308d136 not found: ID does not exist" Oct 14 08:34:10 crc kubenswrapper[5018]: I1014 08:34:10.621013 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" path="/var/lib/kubelet/pods/8954cde6-6c36-40ab-a4fb-d80ed27e4bd8/volumes" Oct 14 08:35:02 crc kubenswrapper[5018]: I1014 08:35:02.463333 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:35:02 crc kubenswrapper[5018]: I1014 08:35:02.464193 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:35:32 crc kubenswrapper[5018]: I1014 08:35:32.464250 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:35:32 crc kubenswrapper[5018]: I1014 08:35:32.465141 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:36:02 crc kubenswrapper[5018]: I1014 08:36:02.464253 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:36:02 crc kubenswrapper[5018]: I1014 08:36:02.464872 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:36:02 crc kubenswrapper[5018]: I1014 08:36:02.464935 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:36:02 crc kubenswrapper[5018]: I1014 08:36:02.465738 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:36:02 crc kubenswrapper[5018]: I1014 08:36:02.465838 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" gracePeriod=600 Oct 14 08:36:02 crc kubenswrapper[5018]: E1014 08:36:02.623537 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:03 crc kubenswrapper[5018]: I1014 08:36:03.493524 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" exitCode=0 Oct 14 08:36:03 crc kubenswrapper[5018]: I1014 08:36:03.493649 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c"} Oct 14 08:36:03 crc kubenswrapper[5018]: I1014 08:36:03.493760 5018 scope.go:117] "RemoveContainer" containerID="818a6ea848d6a4ebf4bdb939ac4b8c3d9a95730398d2a43ac5d0d29d91e5ea99" Oct 14 08:36:03 crc kubenswrapper[5018]: I1014 08:36:03.494537 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:36:03 crc kubenswrapper[5018]: E1014 08:36:03.494994 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:15 crc kubenswrapper[5018]: I1014 08:36:15.605068 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:36:15 crc kubenswrapper[5018]: E1014 08:36:15.606246 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:27 crc kubenswrapper[5018]: I1014 08:36:27.605755 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:36:27 crc kubenswrapper[5018]: E1014 08:36:27.607216 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:41 crc kubenswrapper[5018]: I1014 08:36:41.605196 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:36:41 crc kubenswrapper[5018]: E1014 08:36:41.606320 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.535245 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:36:47 crc kubenswrapper[5018]: E1014 08:36:47.536444 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="extract-content" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.536468 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="extract-content" Oct 14 08:36:47 crc kubenswrapper[5018]: E1014 08:36:47.538253 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="registry-server" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.538292 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="registry-server" Oct 14 08:36:47 crc kubenswrapper[5018]: E1014 08:36:47.538326 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="extract-utilities" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.538345 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="extract-utilities" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.538787 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8954cde6-6c36-40ab-a4fb-d80ed27e4bd8" containerName="registry-server" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.541470 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.551168 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.555257 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.555595 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.555743 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhc2m\" (UniqueName: \"kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.658227 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.658364 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.658402 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhc2m\" (UniqueName: \"kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.659304 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.659386 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.691611 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhc2m\" (UniqueName: \"kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m\") pod \"redhat-marketplace-m496m\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:47 crc kubenswrapper[5018]: I1014 08:36:47.876586 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:48 crc kubenswrapper[5018]: I1014 08:36:48.336888 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:36:49 crc kubenswrapper[5018]: I1014 08:36:49.022834 5018 generic.go:334] "Generic (PLEG): container finished" podID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerID="50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f" exitCode=0 Oct 14 08:36:49 crc kubenswrapper[5018]: I1014 08:36:49.022895 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerDied","Data":"50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f"} Oct 14 08:36:49 crc kubenswrapper[5018]: I1014 08:36:49.022937 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerStarted","Data":"e0484c894cacfcabbc8d97be3a4d8ba19d8c8e9ae2921a23a70de41b56c81136"} Oct 14 08:36:49 crc kubenswrapper[5018]: I1014 08:36:49.027918 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:36:51 crc kubenswrapper[5018]: I1014 08:36:51.047206 5018 generic.go:334] "Generic (PLEG): container finished" podID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerID="f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3" exitCode=0 Oct 14 08:36:51 crc kubenswrapper[5018]: I1014 08:36:51.047995 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerDied","Data":"f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3"} Oct 14 08:36:52 crc kubenswrapper[5018]: I1014 08:36:52.071246 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerStarted","Data":"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94"} Oct 14 08:36:52 crc kubenswrapper[5018]: I1014 08:36:52.094797 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m496m" podStartSLOduration=2.505523096 podStartE2EDuration="5.094775903s" podCreationTimestamp="2025-10-14 08:36:47 +0000 UTC" firstStartedPulling="2025-10-14 08:36:49.027469442 +0000 UTC m=+6425.611516079" lastFinishedPulling="2025-10-14 08:36:51.616722219 +0000 UTC m=+6428.200768886" observedRunningTime="2025-10-14 08:36:52.094173176 +0000 UTC m=+6428.678219803" watchObservedRunningTime="2025-10-14 08:36:52.094775903 +0000 UTC m=+6428.678822560" Oct 14 08:36:53 crc kubenswrapper[5018]: I1014 08:36:53.605729 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:36:53 crc kubenswrapper[5018]: E1014 08:36:53.606089 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:36:57 crc kubenswrapper[5018]: I1014 08:36:57.878728 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:57 crc kubenswrapper[5018]: I1014 08:36:57.879154 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:57 crc kubenswrapper[5018]: I1014 08:36:57.952510 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:58 crc kubenswrapper[5018]: I1014 08:36:58.201722 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:36:58 crc kubenswrapper[5018]: I1014 08:36:58.267371 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.151566 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m496m" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="registry-server" containerID="cri-o://930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94" gracePeriod=2 Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.630245 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.784853 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content\") pod \"3c226804-986e-4bf5-96c0-ab0d74c7d074\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.785018 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhc2m\" (UniqueName: \"kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m\") pod \"3c226804-986e-4bf5-96c0-ab0d74c7d074\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.785086 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities\") pod \"3c226804-986e-4bf5-96c0-ab0d74c7d074\" (UID: \"3c226804-986e-4bf5-96c0-ab0d74c7d074\") " Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.786570 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities" (OuterVolumeSpecName: "utilities") pod "3c226804-986e-4bf5-96c0-ab0d74c7d074" (UID: "3c226804-986e-4bf5-96c0-ab0d74c7d074"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.791436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m" (OuterVolumeSpecName: "kube-api-access-jhc2m") pod "3c226804-986e-4bf5-96c0-ab0d74c7d074" (UID: "3c226804-986e-4bf5-96c0-ab0d74c7d074"). InnerVolumeSpecName "kube-api-access-jhc2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.805509 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c226804-986e-4bf5-96c0-ab0d74c7d074" (UID: "3c226804-986e-4bf5-96c0-ab0d74c7d074"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.887177 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhc2m\" (UniqueName: \"kubernetes.io/projected/3c226804-986e-4bf5-96c0-ab0d74c7d074-kube-api-access-jhc2m\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.887233 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:00 crc kubenswrapper[5018]: I1014 08:37:00.887253 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c226804-986e-4bf5-96c0-ab0d74c7d074-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.166138 5018 generic.go:334] "Generic (PLEG): container finished" podID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerID="930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94" exitCode=0 Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.166191 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerDied","Data":"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94"} Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.166223 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m496m" event={"ID":"3c226804-986e-4bf5-96c0-ab0d74c7d074","Type":"ContainerDied","Data":"e0484c894cacfcabbc8d97be3a4d8ba19d8c8e9ae2921a23a70de41b56c81136"} Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.166242 5018 scope.go:117] "RemoveContainer" containerID="930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.166527 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m496m" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.187592 5018 scope.go:117] "RemoveContainer" containerID="f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.214980 5018 scope.go:117] "RemoveContainer" containerID="50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.278710 5018 scope.go:117] "RemoveContainer" containerID="930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.280514 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:37:01 crc kubenswrapper[5018]: E1014 08:37:01.280551 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94\": container with ID starting with 930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94 not found: ID does not exist" containerID="930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.282047 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94"} err="failed to get container status \"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94\": rpc error: code = NotFound desc = could not find container \"930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94\": container with ID starting with 930d77564274cfd1375e41eef4df18413538ca3762e1be089876325033d1df94 not found: ID does not exist" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.282098 5018 scope.go:117] "RemoveContainer" containerID="f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3" Oct 14 08:37:01 crc kubenswrapper[5018]: E1014 08:37:01.282706 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3\": container with ID starting with f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3 not found: ID does not exist" containerID="f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.282743 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3"} err="failed to get container status \"f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3\": rpc error: code = NotFound desc = could not find container \"f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3\": container with ID starting with f24dd36cf7653d6869239dbffb7b1f5dd60ad9c16e504bf4adbeb132da99aeb3 not found: ID does not exist" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.282774 5018 scope.go:117] "RemoveContainer" containerID="50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f" Oct 14 08:37:01 crc kubenswrapper[5018]: E1014 08:37:01.283230 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f\": container with ID starting with 50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f not found: ID does not exist" containerID="50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.283272 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f"} err="failed to get container status \"50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f\": rpc error: code = NotFound desc = could not find container \"50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f\": container with ID starting with 50b53236a134e3ec9687c406578b9b76f071ecf4793d8098cd7ddd67b64eac6f not found: ID does not exist" Oct 14 08:37:01 crc kubenswrapper[5018]: I1014 08:37:01.290366 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m496m"] Oct 14 08:37:02 crc kubenswrapper[5018]: I1014 08:37:02.624205 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" path="/var/lib/kubelet/pods/3c226804-986e-4bf5-96c0-ab0d74c7d074/volumes" Oct 14 08:37:08 crc kubenswrapper[5018]: I1014 08:37:08.605136 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:37:08 crc kubenswrapper[5018]: E1014 08:37:08.606052 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:37:21 crc kubenswrapper[5018]: I1014 08:37:21.605133 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:37:21 crc kubenswrapper[5018]: E1014 08:37:21.606143 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:37:34 crc kubenswrapper[5018]: I1014 08:37:34.614660 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:37:34 crc kubenswrapper[5018]: E1014 08:37:34.615877 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:37:46 crc kubenswrapper[5018]: I1014 08:37:46.606030 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:37:46 crc kubenswrapper[5018]: E1014 08:37:46.608836 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:37:57 crc kubenswrapper[5018]: I1014 08:37:57.605488 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:37:57 crc kubenswrapper[5018]: E1014 08:37:57.606511 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:38:08 crc kubenswrapper[5018]: I1014 08:38:08.605459 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:38:08 crc kubenswrapper[5018]: E1014 08:38:08.606238 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:38:23 crc kubenswrapper[5018]: I1014 08:38:23.605612 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:38:23 crc kubenswrapper[5018]: E1014 08:38:23.606701 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.360805 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:35 crc kubenswrapper[5018]: E1014 08:38:35.361576 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="registry-server" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.361589 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="registry-server" Oct 14 08:38:35 crc kubenswrapper[5018]: E1014 08:38:35.361608 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="extract-utilities" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.361627 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="extract-utilities" Oct 14 08:38:35 crc kubenswrapper[5018]: E1014 08:38:35.361639 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="extract-content" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.361645 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="extract-content" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.361805 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c226804-986e-4bf5-96c0-ab0d74c7d074" containerName="registry-server" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.362873 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.379755 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.414228 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.414579 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.414744 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtrd\" (UniqueName: \"kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.516190 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.516303 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtrd\" (UniqueName: \"kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.516415 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.516797 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.517380 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.544411 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtrd\" (UniqueName: \"kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd\") pod \"redhat-operators-pmlhx\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.605354 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:38:35 crc kubenswrapper[5018]: E1014 08:38:35.605921 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:38:35 crc kubenswrapper[5018]: I1014 08:38:35.687871 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:36 crc kubenswrapper[5018]: I1014 08:38:36.178107 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:37 crc kubenswrapper[5018]: I1014 08:38:37.186414 5018 generic.go:334] "Generic (PLEG): container finished" podID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerID="5ebb812a12cf700e16e3b15cf5d59f1b0a582e07f6a74aebbc19f8e44e447d3f" exitCode=0 Oct 14 08:38:37 crc kubenswrapper[5018]: I1014 08:38:37.186502 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerDied","Data":"5ebb812a12cf700e16e3b15cf5d59f1b0a582e07f6a74aebbc19f8e44e447d3f"} Oct 14 08:38:37 crc kubenswrapper[5018]: I1014 08:38:37.186859 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerStarted","Data":"5b3588d67b95668ec2c9c1305d022a37dfcb4dfb184b3d5785e87932e7bcd6a9"} Oct 14 08:38:39 crc kubenswrapper[5018]: I1014 08:38:39.212311 5018 generic.go:334] "Generic (PLEG): container finished" podID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerID="33597252b6450496ac18c8cef216fc65b921fc1b69d86c5630188d6acc663600" exitCode=0 Oct 14 08:38:39 crc kubenswrapper[5018]: I1014 08:38:39.212395 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerDied","Data":"33597252b6450496ac18c8cef216fc65b921fc1b69d86c5630188d6acc663600"} Oct 14 08:38:41 crc kubenswrapper[5018]: I1014 08:38:41.243969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerStarted","Data":"bdf26f943508def6e45c75ab334c6c7d838f58f51b07ccbb7991a920499e596d"} Oct 14 08:38:41 crc kubenswrapper[5018]: I1014 08:38:41.278145 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pmlhx" podStartSLOduration=3.333277725 podStartE2EDuration="6.278108601s" podCreationTimestamp="2025-10-14 08:38:35 +0000 UTC" firstStartedPulling="2025-10-14 08:38:37.189811974 +0000 UTC m=+6533.773858611" lastFinishedPulling="2025-10-14 08:38:40.13464281 +0000 UTC m=+6536.718689487" observedRunningTime="2025-10-14 08:38:41.263745501 +0000 UTC m=+6537.847792128" watchObservedRunningTime="2025-10-14 08:38:41.278108601 +0000 UTC m=+6537.862155258" Oct 14 08:38:45 crc kubenswrapper[5018]: I1014 08:38:45.688197 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:45 crc kubenswrapper[5018]: I1014 08:38:45.689113 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:46 crc kubenswrapper[5018]: I1014 08:38:46.751160 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pmlhx" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="registry-server" probeResult="failure" output=< Oct 14 08:38:46 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 08:38:46 crc kubenswrapper[5018]: > Oct 14 08:38:49 crc kubenswrapper[5018]: I1014 08:38:49.604863 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:38:49 crc kubenswrapper[5018]: E1014 08:38:49.605328 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:38:55 crc kubenswrapper[5018]: I1014 08:38:55.777597 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:55 crc kubenswrapper[5018]: I1014 08:38:55.867910 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:56 crc kubenswrapper[5018]: I1014 08:38:56.026367 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:57 crc kubenswrapper[5018]: I1014 08:38:57.417219 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pmlhx" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="registry-server" containerID="cri-o://bdf26f943508def6e45c75ab334c6c7d838f58f51b07ccbb7991a920499e596d" gracePeriod=2 Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.433767 5018 generic.go:334] "Generic (PLEG): container finished" podID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerID="bdf26f943508def6e45c75ab334c6c7d838f58f51b07ccbb7991a920499e596d" exitCode=0 Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.433851 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerDied","Data":"bdf26f943508def6e45c75ab334c6c7d838f58f51b07ccbb7991a920499e596d"} Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.651777 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.779834 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content\") pod \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.779998 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities\") pod \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.780049 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbtrd\" (UniqueName: \"kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd\") pod \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\" (UID: \"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b\") " Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.781457 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities" (OuterVolumeSpecName: "utilities") pod "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" (UID: "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.786097 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd" (OuterVolumeSpecName: "kube-api-access-qbtrd") pod "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" (UID: "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b"). InnerVolumeSpecName "kube-api-access-qbtrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.882359 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.882413 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbtrd\" (UniqueName: \"kubernetes.io/projected/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-kube-api-access-qbtrd\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.883905 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" (UID: "fa16ca70-2a93-4b06-9a21-30fb5ef1db4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:38:58 crc kubenswrapper[5018]: I1014 08:38:58.984497 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.447699 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmlhx" event={"ID":"fa16ca70-2a93-4b06-9a21-30fb5ef1db4b","Type":"ContainerDied","Data":"5b3588d67b95668ec2c9c1305d022a37dfcb4dfb184b3d5785e87932e7bcd6a9"} Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.447785 5018 scope.go:117] "RemoveContainer" containerID="bdf26f943508def6e45c75ab334c6c7d838f58f51b07ccbb7991a920499e596d" Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.447814 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmlhx" Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.483575 5018 scope.go:117] "RemoveContainer" containerID="33597252b6450496ac18c8cef216fc65b921fc1b69d86c5630188d6acc663600" Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.508063 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.517964 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pmlhx"] Oct 14 08:38:59 crc kubenswrapper[5018]: I1014 08:38:59.533596 5018 scope.go:117] "RemoveContainer" containerID="5ebb812a12cf700e16e3b15cf5d59f1b0a582e07f6a74aebbc19f8e44e447d3f" Oct 14 08:39:00 crc kubenswrapper[5018]: I1014 08:39:00.620828 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" path="/var/lib/kubelet/pods/fa16ca70-2a93-4b06-9a21-30fb5ef1db4b/volumes" Oct 14 08:39:03 crc kubenswrapper[5018]: I1014 08:39:03.606178 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:39:03 crc kubenswrapper[5018]: E1014 08:39:03.607277 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:39:06 crc kubenswrapper[5018]: I1014 08:39:06.765062 5018 scope.go:117] "RemoveContainer" containerID="69ff4ba3ad66cda13718843584a3df7cf17247c266293e8d8583732dc0ca161e" Oct 14 08:39:06 crc kubenswrapper[5018]: I1014 08:39:06.788156 5018 scope.go:117] "RemoveContainer" containerID="9b302e8f1cfd864931de5f6f4bec13f42394bd3439340f88cd73d1a0726ab015" Oct 14 08:39:06 crc kubenswrapper[5018]: I1014 08:39:06.845468 5018 scope.go:117] "RemoveContainer" containerID="2a0ba0e528c7064576ebd5cf88b5521b454b3b9271d987d18ec702cc760b8c61" Oct 14 08:39:06 crc kubenswrapper[5018]: I1014 08:39:06.886557 5018 scope.go:117] "RemoveContainer" containerID="3403ddcf9fd9fac69da8b33a0da5faef1eee0bd42b65e205336b4f8ac5366607" Oct 14 08:39:06 crc kubenswrapper[5018]: I1014 08:39:06.918226 5018 scope.go:117] "RemoveContainer" containerID="6e923d496d75ca508f131274a4ffd9d24469d09bdaa1607383b92de1e3c619f7" Oct 14 08:39:16 crc kubenswrapper[5018]: I1014 08:39:16.605838 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:39:16 crc kubenswrapper[5018]: E1014 08:39:16.608156 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:39:29 crc kubenswrapper[5018]: I1014 08:39:29.604935 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:39:29 crc kubenswrapper[5018]: E1014 08:39:29.606018 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:39:44 crc kubenswrapper[5018]: I1014 08:39:44.613305 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:39:44 crc kubenswrapper[5018]: E1014 08:39:44.614063 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:39:57 crc kubenswrapper[5018]: I1014 08:39:57.605942 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:39:57 crc kubenswrapper[5018]: E1014 08:39:57.607149 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.874540 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:01 crc kubenswrapper[5018]: E1014 08:40:01.875790 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="registry-server" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.875825 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="registry-server" Oct 14 08:40:01 crc kubenswrapper[5018]: E1014 08:40:01.875899 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="extract-content" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.875917 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="extract-content" Oct 14 08:40:01 crc kubenswrapper[5018]: E1014 08:40:01.875956 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="extract-utilities" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.875976 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="extract-utilities" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.876341 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa16ca70-2a93-4b06-9a21-30fb5ef1db4b" containerName="registry-server" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.879257 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:01 crc kubenswrapper[5018]: I1014 08:40:01.886059 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.046345 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hngl8\" (UniqueName: \"kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.046420 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.046487 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.148036 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hngl8\" (UniqueName: \"kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.148094 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.148130 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.148869 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.148921 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.181371 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hngl8\" (UniqueName: \"kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8\") pod \"certified-operators-4gwlr\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.205729 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:02 crc kubenswrapper[5018]: I1014 08:40:02.718812 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:03 crc kubenswrapper[5018]: I1014 08:40:03.047967 5018 generic.go:334] "Generic (PLEG): container finished" podID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerID="4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f" exitCode=0 Oct 14 08:40:03 crc kubenswrapper[5018]: I1014 08:40:03.048025 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerDied","Data":"4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f"} Oct 14 08:40:03 crc kubenswrapper[5018]: I1014 08:40:03.048054 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerStarted","Data":"8024aac0e4653a75d4c2e0e5fc309629433df623e0cad6d8c7df6c7a76f945b0"} Oct 14 08:40:05 crc kubenswrapper[5018]: I1014 08:40:05.070741 5018 generic.go:334] "Generic (PLEG): container finished" podID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerID="e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671" exitCode=0 Oct 14 08:40:05 crc kubenswrapper[5018]: I1014 08:40:05.070849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerDied","Data":"e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671"} Oct 14 08:40:06 crc kubenswrapper[5018]: I1014 08:40:06.080144 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerStarted","Data":"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe"} Oct 14 08:40:06 crc kubenswrapper[5018]: I1014 08:40:06.106367 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4gwlr" podStartSLOduration=2.380808557 podStartE2EDuration="5.106344505s" podCreationTimestamp="2025-10-14 08:40:01 +0000 UTC" firstStartedPulling="2025-10-14 08:40:03.049694193 +0000 UTC m=+6619.633740820" lastFinishedPulling="2025-10-14 08:40:05.775230131 +0000 UTC m=+6622.359276768" observedRunningTime="2025-10-14 08:40:06.101918369 +0000 UTC m=+6622.685964996" watchObservedRunningTime="2025-10-14 08:40:06.106344505 +0000 UTC m=+6622.690391122" Oct 14 08:40:07 crc kubenswrapper[5018]: I1014 08:40:07.030889 5018 scope.go:117] "RemoveContainer" containerID="0a6050e9aa1aa2cc5aa7b33655406bfbdb81e224eb7c8413c00f5745b6b373e3" Oct 14 08:40:08 crc kubenswrapper[5018]: I1014 08:40:08.605715 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:40:08 crc kubenswrapper[5018]: E1014 08:40:08.606908 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:40:12 crc kubenswrapper[5018]: I1014 08:40:12.206472 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:12 crc kubenswrapper[5018]: I1014 08:40:12.206966 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:12 crc kubenswrapper[5018]: I1014 08:40:12.266111 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:13 crc kubenswrapper[5018]: I1014 08:40:13.246855 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:13 crc kubenswrapper[5018]: I1014 08:40:13.323653 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.182731 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4gwlr" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="registry-server" containerID="cri-o://c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe" gracePeriod=2 Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.716040 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.814576 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hngl8\" (UniqueName: \"kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8\") pod \"d04596d5-0bf4-4a43-9357-cac7971d6657\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.814716 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content\") pod \"d04596d5-0bf4-4a43-9357-cac7971d6657\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.814798 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities\") pod \"d04596d5-0bf4-4a43-9357-cac7971d6657\" (UID: \"d04596d5-0bf4-4a43-9357-cac7971d6657\") " Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.816113 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities" (OuterVolumeSpecName: "utilities") pod "d04596d5-0bf4-4a43-9357-cac7971d6657" (UID: "d04596d5-0bf4-4a43-9357-cac7971d6657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.816379 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.823657 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8" (OuterVolumeSpecName: "kube-api-access-hngl8") pod "d04596d5-0bf4-4a43-9357-cac7971d6657" (UID: "d04596d5-0bf4-4a43-9357-cac7971d6657"). InnerVolumeSpecName "kube-api-access-hngl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.861064 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d04596d5-0bf4-4a43-9357-cac7971d6657" (UID: "d04596d5-0bf4-4a43-9357-cac7971d6657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.918255 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hngl8\" (UniqueName: \"kubernetes.io/projected/d04596d5-0bf4-4a43-9357-cac7971d6657-kube-api-access-hngl8\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:15 crc kubenswrapper[5018]: I1014 08:40:15.918312 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d04596d5-0bf4-4a43-9357-cac7971d6657-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.194940 5018 generic.go:334] "Generic (PLEG): container finished" podID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerID="c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe" exitCode=0 Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.194999 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerDied","Data":"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe"} Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.195038 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gwlr" event={"ID":"d04596d5-0bf4-4a43-9357-cac7971d6657","Type":"ContainerDied","Data":"8024aac0e4653a75d4c2e0e5fc309629433df623e0cad6d8c7df6c7a76f945b0"} Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.195063 5018 scope.go:117] "RemoveContainer" containerID="c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.195050 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gwlr" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.231187 5018 scope.go:117] "RemoveContainer" containerID="e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.246921 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.251974 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4gwlr"] Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.257312 5018 scope.go:117] "RemoveContainer" containerID="4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.288094 5018 scope.go:117] "RemoveContainer" containerID="c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe" Oct 14 08:40:16 crc kubenswrapper[5018]: E1014 08:40:16.288678 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe\": container with ID starting with c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe not found: ID does not exist" containerID="c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.288729 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe"} err="failed to get container status \"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe\": rpc error: code = NotFound desc = could not find container \"c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe\": container with ID starting with c5e0cb8119a2f132b5bc034fbf64d7f69b8cb5b9b409c97411b8fccdf6226fbe not found: ID does not exist" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.288794 5018 scope.go:117] "RemoveContainer" containerID="e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671" Oct 14 08:40:16 crc kubenswrapper[5018]: E1014 08:40:16.289340 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671\": container with ID starting with e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671 not found: ID does not exist" containerID="e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.289394 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671"} err="failed to get container status \"e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671\": rpc error: code = NotFound desc = could not find container \"e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671\": container with ID starting with e85168c52825c096deb9cdbf8ee05738c22206af63b25741c04b3416f788f671 not found: ID does not exist" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.289430 5018 scope.go:117] "RemoveContainer" containerID="4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f" Oct 14 08:40:16 crc kubenswrapper[5018]: E1014 08:40:16.289925 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f\": container with ID starting with 4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f not found: ID does not exist" containerID="4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.289968 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f"} err="failed to get container status \"4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f\": rpc error: code = NotFound desc = could not find container \"4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f\": container with ID starting with 4179fd046433721e680889f98214753ca4b76b76b33a73a4b8fc061955bb0e6f not found: ID does not exist" Oct 14 08:40:16 crc kubenswrapper[5018]: I1014 08:40:16.623897 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" path="/var/lib/kubelet/pods/d04596d5-0bf4-4a43-9357-cac7971d6657/volumes" Oct 14 08:40:22 crc kubenswrapper[5018]: I1014 08:40:22.605525 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:40:22 crc kubenswrapper[5018]: E1014 08:40:22.606296 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:40:33 crc kubenswrapper[5018]: I1014 08:40:33.605653 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:40:33 crc kubenswrapper[5018]: E1014 08:40:33.606706 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.508601 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:40:40 crc kubenswrapper[5018]: E1014 08:40:40.510010 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="extract-content" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.510044 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="extract-content" Oct 14 08:40:40 crc kubenswrapper[5018]: E1014 08:40:40.510083 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="extract-utilities" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.510100 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="extract-utilities" Oct 14 08:40:40 crc kubenswrapper[5018]: E1014 08:40:40.510151 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="registry-server" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.510169 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="registry-server" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.510565 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04596d5-0bf4-4a43-9357-cac7971d6657" containerName="registry-server" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.511787 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.514676 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-k6crk" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.523750 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.622453 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzv95\" (UniqueName: \"kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.622777 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.724973 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.725068 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzv95\" (UniqueName: \"kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.729067 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.729151 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3f360520ff0ad80da1bafa7748da52808c315279775d04d1ee747205ed8e4e86/globalmount\"" pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.767447 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzv95\" (UniqueName: \"kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.776421 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") pod \"mariadb-copy-data\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " pod="openstack/mariadb-copy-data" Oct 14 08:40:40 crc kubenswrapper[5018]: I1014 08:40:40.847494 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 08:40:41 crc kubenswrapper[5018]: I1014 08:40:41.462951 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:40:41 crc kubenswrapper[5018]: W1014 08:40:41.468149 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35063fdb_abe2_4d3a_8833_8d62162a1a01.slice/crio-a9b712927c6c2a994d8f40700c7dd0ed2da3263737850e387fd081507abaa976 WatchSource:0}: Error finding container a9b712927c6c2a994d8f40700c7dd0ed2da3263737850e387fd081507abaa976: Status 404 returned error can't find the container with id a9b712927c6c2a994d8f40700c7dd0ed2da3263737850e387fd081507abaa976 Oct 14 08:40:42 crc kubenswrapper[5018]: I1014 08:40:42.473848 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"35063fdb-abe2-4d3a-8833-8d62162a1a01","Type":"ContainerStarted","Data":"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c"} Oct 14 08:40:42 crc kubenswrapper[5018]: I1014 08:40:42.474339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"35063fdb-abe2-4d3a-8833-8d62162a1a01","Type":"ContainerStarted","Data":"a9b712927c6c2a994d8f40700c7dd0ed2da3263737850e387fd081507abaa976"} Oct 14 08:40:42 crc kubenswrapper[5018]: I1014 08:40:42.504730 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.504696847 podStartE2EDuration="3.504696847s" podCreationTimestamp="2025-10-14 08:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:40:42.49601311 +0000 UTC m=+6659.080059767" watchObservedRunningTime="2025-10-14 08:40:42.504696847 +0000 UTC m=+6659.088743484" Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.237131 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.238827 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.248187 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.416793 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4br5\" (UniqueName: \"kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5\") pod \"mariadb-client\" (UID: \"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42\") " pod="openstack/mariadb-client" Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.519758 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4br5\" (UniqueName: \"kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5\") pod \"mariadb-client\" (UID: \"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42\") " pod="openstack/mariadb-client" Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.563611 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4br5\" (UniqueName: \"kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5\") pod \"mariadb-client\" (UID: \"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42\") " pod="openstack/mariadb-client" Oct 14 08:40:45 crc kubenswrapper[5018]: I1014 08:40:45.574207 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:46 crc kubenswrapper[5018]: I1014 08:40:46.137656 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:46 crc kubenswrapper[5018]: W1014 08:40:46.150883 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb4acdc8_f88a_4f8f_a5f4_576a1c56bf42.slice/crio-005006c1325fec957e92f90f22270f71edcea8140bfa21979e58d202e68ef158 WatchSource:0}: Error finding container 005006c1325fec957e92f90f22270f71edcea8140bfa21979e58d202e68ef158: Status 404 returned error can't find the container with id 005006c1325fec957e92f90f22270f71edcea8140bfa21979e58d202e68ef158 Oct 14 08:40:46 crc kubenswrapper[5018]: I1014 08:40:46.535010 5018 generic.go:334] "Generic (PLEG): container finished" podID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" containerID="10f130c9b8f9466bd0e5944aaa5001f9528abd0308435ffa589e4683d51b114b" exitCode=0 Oct 14 08:40:46 crc kubenswrapper[5018]: I1014 08:40:46.535090 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42","Type":"ContainerDied","Data":"10f130c9b8f9466bd0e5944aaa5001f9528abd0308435ffa589e4683d51b114b"} Oct 14 08:40:46 crc kubenswrapper[5018]: I1014 08:40:46.535169 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42","Type":"ContainerStarted","Data":"005006c1325fec957e92f90f22270f71edcea8140bfa21979e58d202e68ef158"} Oct 14 08:40:47 crc kubenswrapper[5018]: I1014 08:40:47.950916 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:47 crc kubenswrapper[5018]: I1014 08:40:47.983573 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42/mariadb-client/0.log" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.011825 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.016732 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.067072 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4br5\" (UniqueName: \"kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5\") pod \"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42\" (UID: \"bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42\") " Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.073901 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5" (OuterVolumeSpecName: "kube-api-access-m4br5") pod "bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" (UID: "bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42"). InnerVolumeSpecName "kube-api-access-m4br5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.166321 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:48 crc kubenswrapper[5018]: E1014 08:40:48.166694 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" containerName="mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.166717 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" containerName="mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.166897 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" containerName="mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.167459 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.169341 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4br5\" (UniqueName: \"kubernetes.io/projected/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42-kube-api-access-m4br5\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.183582 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.271188 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2cnk\" (UniqueName: \"kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk\") pod \"mariadb-client\" (UID: \"d4b174e4-3c36-4ce7-8c61-804cc3650926\") " pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.373090 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2cnk\" (UniqueName: \"kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk\") pod \"mariadb-client\" (UID: \"d4b174e4-3c36-4ce7-8c61-804cc3650926\") " pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.402989 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2cnk\" (UniqueName: \"kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk\") pod \"mariadb-client\" (UID: \"d4b174e4-3c36-4ce7-8c61-804cc3650926\") " pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.500669 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.554716 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="005006c1325fec957e92f90f22270f71edcea8140bfa21979e58d202e68ef158" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.554970 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.594530 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" podUID="d4b174e4-3c36-4ce7-8c61-804cc3650926" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.605221 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:40:48 crc kubenswrapper[5018]: E1014 08:40:48.605640 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.642028 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42" path="/var/lib/kubelet/pods/bb4acdc8-f88a-4f8f-a5f4-576a1c56bf42/volumes" Oct 14 08:40:48 crc kubenswrapper[5018]: I1014 08:40:48.793419 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:48 crc kubenswrapper[5018]: W1014 08:40:48.801064 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4b174e4_3c36_4ce7_8c61_804cc3650926.slice/crio-c211ae78e5685562f0a050792cc45d4a864f3b3c3f0c2628bfaed1a4df5f06a8 WatchSource:0}: Error finding container c211ae78e5685562f0a050792cc45d4a864f3b3c3f0c2628bfaed1a4df5f06a8: Status 404 returned error can't find the container with id c211ae78e5685562f0a050792cc45d4a864f3b3c3f0c2628bfaed1a4df5f06a8 Oct 14 08:40:49 crc kubenswrapper[5018]: I1014 08:40:49.573439 5018 generic.go:334] "Generic (PLEG): container finished" podID="d4b174e4-3c36-4ce7-8c61-804cc3650926" containerID="d2bd12da785c7782d1bcd148149e961b39f28373632a498c2c01b8d852f3ec54" exitCode=0 Oct 14 08:40:49 crc kubenswrapper[5018]: I1014 08:40:49.573503 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"d4b174e4-3c36-4ce7-8c61-804cc3650926","Type":"ContainerDied","Data":"d2bd12da785c7782d1bcd148149e961b39f28373632a498c2c01b8d852f3ec54"} Oct 14 08:40:49 crc kubenswrapper[5018]: I1014 08:40:49.573545 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"d4b174e4-3c36-4ce7-8c61-804cc3650926","Type":"ContainerStarted","Data":"c211ae78e5685562f0a050792cc45d4a864f3b3c3f0c2628bfaed1a4df5f06a8"} Oct 14 08:40:50 crc kubenswrapper[5018]: I1014 08:40:50.950946 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:50 crc kubenswrapper[5018]: I1014 08:40:50.976883 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_d4b174e4-3c36-4ce7-8c61-804cc3650926/mariadb-client/0.log" Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.011917 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.021238 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.123313 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2cnk\" (UniqueName: \"kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk\") pod \"d4b174e4-3c36-4ce7-8c61-804cc3650926\" (UID: \"d4b174e4-3c36-4ce7-8c61-804cc3650926\") " Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.139654 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk" (OuterVolumeSpecName: "kube-api-access-g2cnk") pod "d4b174e4-3c36-4ce7-8c61-804cc3650926" (UID: "d4b174e4-3c36-4ce7-8c61-804cc3650926"). InnerVolumeSpecName "kube-api-access-g2cnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.226153 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2cnk\" (UniqueName: \"kubernetes.io/projected/d4b174e4-3c36-4ce7-8c61-804cc3650926-kube-api-access-g2cnk\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.599190 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c211ae78e5685562f0a050792cc45d4a864f3b3c3f0c2628bfaed1a4df5f06a8" Oct 14 08:40:51 crc kubenswrapper[5018]: I1014 08:40:51.599260 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:40:52 crc kubenswrapper[5018]: I1014 08:40:52.631293 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4b174e4-3c36-4ce7-8c61-804cc3650926" path="/var/lib/kubelet/pods/d4b174e4-3c36-4ce7-8c61-804cc3650926/volumes" Oct 14 08:41:03 crc kubenswrapper[5018]: I1014 08:41:03.605053 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:41:04 crc kubenswrapper[5018]: I1014 08:41:04.728058 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee"} Oct 14 08:41:21 crc kubenswrapper[5018]: E1014 08:41:21.711307 5018 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.159:52630->38.102.83.159:32939: write tcp 38.102.83.159:52630->38.102.83.159:32939: write: connection reset by peer Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.526131 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:41:26 crc kubenswrapper[5018]: E1014 08:41:26.527153 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b174e4-3c36-4ce7-8c61-804cc3650926" containerName="mariadb-client" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.527180 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b174e4-3c36-4ce7-8c61-804cc3650926" containerName="mariadb-client" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.527501 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b174e4-3c36-4ce7-8c61-804cc3650926" containerName="mariadb-client" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.528970 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.533121 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.534497 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.535037 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.535207 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-x2q9g" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.536508 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.543697 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.552610 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.555889 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.559366 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.561386 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.583481 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.594539 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.644892 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.644990 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645022 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645053 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645073 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645096 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645135 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645194 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.645294 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9tr\" (UniqueName: \"kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.646317 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648238 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdvt\" (UniqueName: \"kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648382 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648497 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648733 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.648765 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750061 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750116 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750154 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750191 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9tr\" (UniqueName: \"kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750225 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750268 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fqf2\" (UniqueName: \"kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750296 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750327 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cdvt\" (UniqueName: \"kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750380 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750404 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750441 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750470 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750497 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750527 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750555 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750586 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750642 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750670 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750700 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750725 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750759 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750783 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750808 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750842 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.750977 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.751817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.753645 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.754300 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.755207 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.756208 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.757147 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.757187 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/296ba3a485fd340ffb807b64fd47cbaae28f6d2723c3db095db4889ad1ad7d53/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.757306 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.757337 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5c0f22dd51a4c18094de15ff4428af2c84da1afc423257a98074aaf811daff3a/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.758984 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.761597 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.762569 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.763280 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.763332 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.779383 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.779512 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9tr\" (UniqueName: \"kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.784408 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cdvt\" (UniqueName: \"kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.792603 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"ovsdbserver-nb-0\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.799460 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") pod \"ovsdbserver-nb-2\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.851750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853274 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853425 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fqf2\" (UniqueName: \"kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853533 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853201 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853722 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.853991 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.854469 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.854600 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.854743 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.855608 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.857311 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.857654 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.857691 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/48918289118c61a7a75769894fadf599f4e453a1ea0cbd4ea48bf3f3b0383054/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.859431 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.862103 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.871980 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fqf2\" (UniqueName: \"kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.872600 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.894461 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") pod \"ovsdbserver-nb-1\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.903029 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:26 crc kubenswrapper[5018]: I1014 08:41:26.919008 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:27 crc kubenswrapper[5018]: I1014 08:41:27.560137 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:41:27 crc kubenswrapper[5018]: W1014 08:41:27.570082 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb67bc93c_1c02_47fa_b2a5_57560256e89b.slice/crio-0bd72932314fd0730fac8cd026bc5e129bd1070837a76c295efa99938592dd2d WatchSource:0}: Error finding container 0bd72932314fd0730fac8cd026bc5e129bd1070837a76c295efa99938592dd2d: Status 404 returned error can't find the container with id 0bd72932314fd0730fac8cd026bc5e129bd1070837a76c295efa99938592dd2d Oct 14 08:41:27 crc kubenswrapper[5018]: I1014 08:41:27.650060 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:41:27 crc kubenswrapper[5018]: W1014 08:41:27.666555 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcdada18_a44b_4d39_b4f6_2bc9fef4530c.slice/crio-f127d2b035406ce1d2e06959048355959f5819539681b83eac1c3a7ba3864e9b WatchSource:0}: Error finding container f127d2b035406ce1d2e06959048355959f5819539681b83eac1c3a7ba3864e9b: Status 404 returned error can't find the container with id f127d2b035406ce1d2e06959048355959f5819539681b83eac1c3a7ba3864e9b Oct 14 08:41:27 crc kubenswrapper[5018]: I1014 08:41:27.982747 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerStarted","Data":"0bd72932314fd0730fac8cd026bc5e129bd1070837a76c295efa99938592dd2d"} Oct 14 08:41:27 crc kubenswrapper[5018]: I1014 08:41:27.986050 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerStarted","Data":"f127d2b035406ce1d2e06959048355959f5819539681b83eac1c3a7ba3864e9b"} Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.073851 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.075422 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.084512 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-w5bz4" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.085725 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.086011 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.086254 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.097220 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.120092 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.122428 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.126087 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.128531 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.135988 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.140809 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.181932 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.181982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182040 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182065 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182096 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182121 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182252 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182296 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182328 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182398 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182431 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbrg7\" (UniqueName: \"kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182473 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182510 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182542 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182576 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182639 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182675 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182711 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182733 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n977\" (UniqueName: \"kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182774 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182800 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182838 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.182866 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5lp\" (UniqueName: \"kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.183145 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285269 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285326 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285358 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285387 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285414 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285440 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285469 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n977\" (UniqueName: \"kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285487 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285513 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285531 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285557 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285578 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5lp\" (UniqueName: \"kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285596 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285655 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285677 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285721 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285745 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285768 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285791 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285819 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285838 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285857 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285880 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.285898 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbrg7\" (UniqueName: \"kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.286267 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.286313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.286759 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.287327 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.287814 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.288920 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.290691 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.291005 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.291138 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.291277 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.291607 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295294 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295380 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295413 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6233fae2600b34c8175968d92f4e0f3446b5af046a76461acad1979fd0637f1b/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295419 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295441 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295763 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.295881 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8395c6fe2d90fc274251f8753421d859f1940c089b0f29c152f3473099639a7f/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.296742 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.301125 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.301220 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8e91c4a863bd4b5cfe31ce299341bf7e453582f19f0d110857900037fe7b5be5/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.302353 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.304265 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.306306 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n977\" (UniqueName: \"kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.306595 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.307583 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbrg7\" (UniqueName: \"kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.307942 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5lp\" (UniqueName: \"kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.333759 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") pod \"ovsdbserver-sb-0\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.333990 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") pod \"ovsdbserver-sb-2\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.333997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") pod \"ovsdbserver-sb-1\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.415874 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.445762 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.457514 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.688059 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:41:28 crc kubenswrapper[5018]: W1014 08:41:28.696879 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod495a63bc_82b8_4a2b_83ab_5aed23f6f493.slice/crio-64b4317a8e03b1fdb07ea4956329cfde29ff6af5f99d87eb57b15f3b32f4974f WatchSource:0}: Error finding container 64b4317a8e03b1fdb07ea4956329cfde29ff6af5f99d87eb57b15f3b32f4974f: Status 404 returned error can't find the container with id 64b4317a8e03b1fdb07ea4956329cfde29ff6af5f99d87eb57b15f3b32f4974f Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.969717 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:41:28 crc kubenswrapper[5018]: W1014 08:41:28.984579 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9703a145_2c14_4a43_a767_e6def8fe6063.slice/crio-8f33b654947a44e55ec8833c78e946654d9bb0accc7e767d8e0ef9ab5ac39e41 WatchSource:0}: Error finding container 8f33b654947a44e55ec8833c78e946654d9bb0accc7e767d8e0ef9ab5ac39e41: Status 404 returned error can't find the container with id 8f33b654947a44e55ec8833c78e946654d9bb0accc7e767d8e0ef9ab5ac39e41 Oct 14 08:41:28 crc kubenswrapper[5018]: I1014 08:41:28.995971 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerStarted","Data":"64b4317a8e03b1fdb07ea4956329cfde29ff6af5f99d87eb57b15f3b32f4974f"} Oct 14 08:41:29 crc kubenswrapper[5018]: W1014 08:41:29.058412 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a2d465_d49e_4f9d_a1c9_d5f674ae30e9.slice/crio-5c651b3c84b92683765b1a9ea0862b522bf4b38c23c9ae7ee3e86fdefce2de3e WatchSource:0}: Error finding container 5c651b3c84b92683765b1a9ea0862b522bf4b38c23c9ae7ee3e86fdefce2de3e: Status 404 returned error can't find the container with id 5c651b3c84b92683765b1a9ea0862b522bf4b38c23c9ae7ee3e86fdefce2de3e Oct 14 08:41:29 crc kubenswrapper[5018]: I1014 08:41:29.058547 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:41:29 crc kubenswrapper[5018]: I1014 08:41:29.589911 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:41:30 crc kubenswrapper[5018]: I1014 08:41:30.006076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerStarted","Data":"5c651b3c84b92683765b1a9ea0862b522bf4b38c23c9ae7ee3e86fdefce2de3e"} Oct 14 08:41:30 crc kubenswrapper[5018]: I1014 08:41:30.007456 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerStarted","Data":"d77a1d281e4fe33ee612e40d06af4451ebeb1d0d1001be5ccb6285dd06990598"} Oct 14 08:41:30 crc kubenswrapper[5018]: I1014 08:41:30.008611 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerStarted","Data":"8f33b654947a44e55ec8833c78e946654d9bb0accc7e767d8e0ef9ab5ac39e41"} Oct 14 08:41:32 crc kubenswrapper[5018]: I1014 08:41:32.027287 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerStarted","Data":"2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad"} Oct 14 08:41:32 crc kubenswrapper[5018]: I1014 08:41:32.031840 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerStarted","Data":"f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a"} Oct 14 08:41:32 crc kubenswrapper[5018]: I1014 08:41:32.033604 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerStarted","Data":"6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.045566 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerStarted","Data":"8518cc00808c52898e754e85cacd8523a46b73a34084d59361518ea964572758"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.048871 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerStarted","Data":"19de2087aaf37731d242d90edcb3f4e4a539c73a836422704d5ea34ead7356dc"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.048928 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerStarted","Data":"6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.050670 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerStarted","Data":"b8af3ead5cf9c4d299de3975f6ab77037e7e1da24c12c876f67ca25674008762"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.050813 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerStarted","Data":"fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.054380 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerStarted","Data":"e06750d462802f2f98ea3875ae3f637ed5b88aae6f70ae700e5336271d04d93c"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.057428 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerStarted","Data":"f0b29be9a0afeb5e7192333db87e868120fdaf92944e86c96054f44887940f2f"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.059805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerStarted","Data":"f5a9062ecbff01d430f8c45450a3ea8dfae4ad161b221c5afda3f75ccde512b0"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.059956 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerStarted","Data":"6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8"} Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.068834 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=5.407233184 podStartE2EDuration="8.068821449s" podCreationTimestamp="2025-10-14 08:41:25 +0000 UTC" firstStartedPulling="2025-10-14 08:41:28.700407732 +0000 UTC m=+6705.284454359" lastFinishedPulling="2025-10-14 08:41:31.361995997 +0000 UTC m=+6707.946042624" observedRunningTime="2025-10-14 08:41:33.064916958 +0000 UTC m=+6709.648963585" watchObservedRunningTime="2025-10-14 08:41:33.068821449 +0000 UTC m=+6709.652868076" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.097153 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=2.625179222 podStartE2EDuration="6.097127096s" podCreationTimestamp="2025-10-14 08:41:27 +0000 UTC" firstStartedPulling="2025-10-14 08:41:29.061054448 +0000 UTC m=+6705.645101075" lastFinishedPulling="2025-10-14 08:41:32.533002322 +0000 UTC m=+6709.117048949" observedRunningTime="2025-10-14 08:41:33.091106374 +0000 UTC m=+6709.675153001" watchObservedRunningTime="2025-10-14 08:41:33.097127096 +0000 UTC m=+6709.681173733" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.118209 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.378618236 podStartE2EDuration="8.118191056s" podCreationTimestamp="2025-10-14 08:41:25 +0000 UTC" firstStartedPulling="2025-10-14 08:41:27.573088141 +0000 UTC m=+6704.157134808" lastFinishedPulling="2025-10-14 08:41:31.312661001 +0000 UTC m=+6707.896707628" observedRunningTime="2025-10-14 08:41:33.115116478 +0000 UTC m=+6709.699163115" watchObservedRunningTime="2025-10-14 08:41:33.118191056 +0000 UTC m=+6709.702237693" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.162146 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.505317936 podStartE2EDuration="8.162109987s" podCreationTimestamp="2025-10-14 08:41:25 +0000 UTC" firstStartedPulling="2025-10-14 08:41:27.670011133 +0000 UTC m=+6704.254057790" lastFinishedPulling="2025-10-14 08:41:31.326803214 +0000 UTC m=+6707.910849841" observedRunningTime="2025-10-14 08:41:33.137416804 +0000 UTC m=+6709.721463451" watchObservedRunningTime="2025-10-14 08:41:33.162109987 +0000 UTC m=+6709.746156644" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.164299 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.196127499 podStartE2EDuration="6.164281559s" podCreationTimestamp="2025-10-14 08:41:27 +0000 UTC" firstStartedPulling="2025-10-14 08:41:28.991804955 +0000 UTC m=+6705.575851582" lastFinishedPulling="2025-10-14 08:41:31.959959005 +0000 UTC m=+6708.544005642" observedRunningTime="2025-10-14 08:41:33.154868601 +0000 UTC m=+6709.738915228" watchObservedRunningTime="2025-10-14 08:41:33.164281559 +0000 UTC m=+6709.748328226" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.190229 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.770420322 podStartE2EDuration="6.190212108s" podCreationTimestamp="2025-10-14 08:41:27 +0000 UTC" firstStartedPulling="2025-10-14 08:41:29.587387624 +0000 UTC m=+6706.171434251" lastFinishedPulling="2025-10-14 08:41:32.00717941 +0000 UTC m=+6708.591226037" observedRunningTime="2025-10-14 08:41:33.182878879 +0000 UTC m=+6709.766925526" watchObservedRunningTime="2025-10-14 08:41:33.190212108 +0000 UTC m=+6709.774258745" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.416689 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.445942 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:33 crc kubenswrapper[5018]: I1014 08:41:33.458655 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:34 crc kubenswrapper[5018]: I1014 08:41:34.416768 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:34 crc kubenswrapper[5018]: I1014 08:41:34.446107 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:34 crc kubenswrapper[5018]: I1014 08:41:34.458149 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:35 crc kubenswrapper[5018]: I1014 08:41:35.873690 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:35 crc kubenswrapper[5018]: I1014 08:41:35.904372 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:35 crc kubenswrapper[5018]: I1014 08:41:35.919816 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:35 crc kubenswrapper[5018]: I1014 08:41:35.957601 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:35 crc kubenswrapper[5018]: I1014 08:41:35.981430 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.001001 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.094155 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.094214 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.094231 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.144293 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.151566 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.157420 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.447313 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.448703 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.450859 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.469966 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.552027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.552117 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.552229 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v44fd\" (UniqueName: \"kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.552405 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.654540 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.654663 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.654744 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.654766 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v44fd\" (UniqueName: \"kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.658466 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.658669 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.658880 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.692147 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v44fd\" (UniqueName: \"kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd\") pod \"dnsmasq-dns-5cffbd857f-4475k\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:36 crc kubenswrapper[5018]: I1014 08:41:36.766019 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.254473 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.500079 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.521764 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.537959 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.584807 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.611851 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.611905 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.786044 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.813637 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.822319 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.824025 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.834271 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.977700 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.977893 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.978013 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmq4k\" (UniqueName: \"kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.978082 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:37 crc kubenswrapper[5018]: I1014 08:41:37.978122 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.079923 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.079999 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmq4k\" (UniqueName: \"kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.080031 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.080055 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.080121 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.080815 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.080838 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.081382 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.081459 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.105391 5018 generic.go:334] "Generic (PLEG): container finished" podID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerID="cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e" exitCode=0 Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.105498 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" event={"ID":"5b9a2518-de8e-4966-a69f-7e0430138e07","Type":"ContainerDied","Data":"cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e"} Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.105549 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" event={"ID":"5b9a2518-de8e-4966-a69f-7e0430138e07","Type":"ContainerStarted","Data":"5a5d5ca8469844a0f653a2319c1bbc61d08e7b228d2acbad8ad17319c3eaa2ff"} Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.113051 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmq4k\" (UniqueName: \"kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k\") pod \"dnsmasq-dns-5bc8959bc9-cn9t4\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.143767 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:38 crc kubenswrapper[5018]: I1014 08:41:38.598103 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:41:38 crc kubenswrapper[5018]: W1014 08:41:38.599285 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod919e2736_0848_4179_a50f_23782208a439.slice/crio-339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649 WatchSource:0}: Error finding container 339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649: Status 404 returned error can't find the container with id 339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649 Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.120915 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" event={"ID":"5b9a2518-de8e-4966-a69f-7e0430138e07","Type":"ContainerStarted","Data":"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2"} Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.121056 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="dnsmasq-dns" containerID="cri-o://62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2" gracePeriod=10 Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.121521 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.129388 5018 generic.go:334] "Generic (PLEG): container finished" podID="919e2736-0848-4179-a50f-23782208a439" containerID="ad988727da44fbb625c5d40108ba5ec483d18f15f3181e6860c2a9f5b2ae2707" exitCode=0 Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.129479 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" event={"ID":"919e2736-0848-4179-a50f-23782208a439","Type":"ContainerDied","Data":"ad988727da44fbb625c5d40108ba5ec483d18f15f3181e6860c2a9f5b2ae2707"} Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.129567 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" event={"ID":"919e2736-0848-4179-a50f-23782208a439","Type":"ContainerStarted","Data":"339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649"} Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.181067 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" podStartSLOduration=3.180356332 podStartE2EDuration="3.180356332s" podCreationTimestamp="2025-10-14 08:41:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:41:39.156975126 +0000 UTC m=+6715.741021823" watchObservedRunningTime="2025-10-14 08:41:39.180356332 +0000 UTC m=+6715.764402979" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.586358 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.717535 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb\") pod \"5b9a2518-de8e-4966-a69f-7e0430138e07\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.718137 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v44fd\" (UniqueName: \"kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd\") pod \"5b9a2518-de8e-4966-a69f-7e0430138e07\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.718260 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc\") pod \"5b9a2518-de8e-4966-a69f-7e0430138e07\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.718511 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config\") pod \"5b9a2518-de8e-4966-a69f-7e0430138e07\" (UID: \"5b9a2518-de8e-4966-a69f-7e0430138e07\") " Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.723102 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd" (OuterVolumeSpecName: "kube-api-access-v44fd") pod "5b9a2518-de8e-4966-a69f-7e0430138e07" (UID: "5b9a2518-de8e-4966-a69f-7e0430138e07"). InnerVolumeSpecName "kube-api-access-v44fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.773743 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config" (OuterVolumeSpecName: "config") pod "5b9a2518-de8e-4966-a69f-7e0430138e07" (UID: "5b9a2518-de8e-4966-a69f-7e0430138e07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.774605 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b9a2518-de8e-4966-a69f-7e0430138e07" (UID: "5b9a2518-de8e-4966-a69f-7e0430138e07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.775371 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b9a2518-de8e-4966-a69f-7e0430138e07" (UID: "5b9a2518-de8e-4966-a69f-7e0430138e07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.822073 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.822128 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v44fd\" (UniqueName: \"kubernetes.io/projected/5b9a2518-de8e-4966-a69f-7e0430138e07-kube-api-access-v44fd\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.822151 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:39 crc kubenswrapper[5018]: I1014 08:41:39.822170 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b9a2518-de8e-4966-a69f-7e0430138e07-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.154696 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.155150 5018 generic.go:334] "Generic (PLEG): container finished" podID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerID="62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2" exitCode=0 Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.155408 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" Oct 14 08:41:40 crc kubenswrapper[5018]: E1014 08:41:40.155501 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="init" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.155694 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="init" Oct 14 08:41:40 crc kubenswrapper[5018]: E1014 08:41:40.155760 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="dnsmasq-dns" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.155778 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="dnsmasq-dns" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.156819 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" containerName="dnsmasq-dns" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.157898 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" event={"ID":"5b9a2518-de8e-4966-a69f-7e0430138e07","Type":"ContainerDied","Data":"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2"} Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.157955 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cffbd857f-4475k" event={"ID":"5b9a2518-de8e-4966-a69f-7e0430138e07","Type":"ContainerDied","Data":"5a5d5ca8469844a0f653a2319c1bbc61d08e7b228d2acbad8ad17319c3eaa2ff"} Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.157988 5018 scope.go:117] "RemoveContainer" containerID="62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.158185 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.181601 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.202607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" event={"ID":"919e2736-0848-4179-a50f-23782208a439","Type":"ContainerStarted","Data":"6ec0fca037768af2c6ab54de01240b5c1ec59460aa82d49fee10b48b4e0cd173"} Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.202821 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.202826 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.232261 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fpkw\" (UniqueName: \"kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.232309 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.232377 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.245919 5018 scope.go:117] "RemoveContainer" containerID="cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.258418 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.264062 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cffbd857f-4475k"] Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.265085 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" podStartSLOduration=3.265068139 podStartE2EDuration="3.265068139s" podCreationTimestamp="2025-10-14 08:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:41:40.257427881 +0000 UTC m=+6716.841474548" watchObservedRunningTime="2025-10-14 08:41:40.265068139 +0000 UTC m=+6716.849114766" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.281270 5018 scope.go:117] "RemoveContainer" containerID="62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2" Oct 14 08:41:40 crc kubenswrapper[5018]: E1014 08:41:40.281790 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2\": container with ID starting with 62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2 not found: ID does not exist" containerID="62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.281830 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2"} err="failed to get container status \"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2\": rpc error: code = NotFound desc = could not find container \"62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2\": container with ID starting with 62e7085b441427a52b61df0a01fb40719eebc25e38affda173ddbe79a4e461e2 not found: ID does not exist" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.281854 5018 scope.go:117] "RemoveContainer" containerID="cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e" Oct 14 08:41:40 crc kubenswrapper[5018]: E1014 08:41:40.282329 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e\": container with ID starting with cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e not found: ID does not exist" containerID="cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.282354 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e"} err="failed to get container status \"cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e\": rpc error: code = NotFound desc = could not find container \"cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e\": container with ID starting with cfc0ed134d52b7d015f824ad974b45cbfadadaa8693e4f34b3c7fc2cc651037e not found: ID does not exist" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.339874 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fpkw\" (UniqueName: \"kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.339948 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.340053 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.347194 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.347511 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/85dbba56b7b9f31e11f05f0fc6d4399b4d3ddab086d4620f93a02b423ebd2d87/globalmount\"" pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.348894 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.357887 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fpkw\" (UniqueName: \"kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.394883 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") pod \"ovn-copy-data\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.540808 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.618243 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9a2518-de8e-4966-a69f-7e0430138e07" path="/var/lib/kubelet/pods/5b9a2518-de8e-4966-a69f-7e0430138e07/volumes" Oct 14 08:41:40 crc kubenswrapper[5018]: I1014 08:41:40.918357 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:41:40 crc kubenswrapper[5018]: W1014 08:41:40.923011 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43c4ab03_dd63_4001_9a55_833b15bc11fc.slice/crio-7cc99a0f83af0d084556ee8f65b92f6ed60d9a41a2cd3497f97f3da4f22ac714 WatchSource:0}: Error finding container 7cc99a0f83af0d084556ee8f65b92f6ed60d9a41a2cd3497f97f3da4f22ac714: Status 404 returned error can't find the container with id 7cc99a0f83af0d084556ee8f65b92f6ed60d9a41a2cd3497f97f3da4f22ac714 Oct 14 08:41:41 crc kubenswrapper[5018]: I1014 08:41:41.212606 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"43c4ab03-dd63-4001-9a55-833b15bc11fc","Type":"ContainerStarted","Data":"7cc99a0f83af0d084556ee8f65b92f6ed60d9a41a2cd3497f97f3da4f22ac714"} Oct 14 08:41:42 crc kubenswrapper[5018]: I1014 08:41:42.229190 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"43c4ab03-dd63-4001-9a55-833b15bc11fc","Type":"ContainerStarted","Data":"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18"} Oct 14 08:41:42 crc kubenswrapper[5018]: I1014 08:41:42.266384 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.039826135 podStartE2EDuration="3.26635347s" podCreationTimestamp="2025-10-14 08:41:39 +0000 UTC" firstStartedPulling="2025-10-14 08:41:40.926156455 +0000 UTC m=+6717.510203092" lastFinishedPulling="2025-10-14 08:41:41.15268379 +0000 UTC m=+6717.736730427" observedRunningTime="2025-10-14 08:41:42.252707481 +0000 UTC m=+6718.836754148" watchObservedRunningTime="2025-10-14 08:41:42.26635347 +0000 UTC m=+6718.850400127" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.145899 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.232880 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.233219 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="dnsmasq-dns" containerID="cri-o://578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde" gracePeriod=10 Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.328709 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.3:5353: connect: connection refused" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.693340 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.798900 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config\") pod \"9383f078-40c5-48ab-937b-31593364ecc7\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.798967 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljwmr\" (UniqueName: \"kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr\") pod \"9383f078-40c5-48ab-937b-31593364ecc7\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.799084 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc\") pod \"9383f078-40c5-48ab-937b-31593364ecc7\" (UID: \"9383f078-40c5-48ab-937b-31593364ecc7\") " Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.805944 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr" (OuterVolumeSpecName: "kube-api-access-ljwmr") pod "9383f078-40c5-48ab-937b-31593364ecc7" (UID: "9383f078-40c5-48ab-937b-31593364ecc7"). InnerVolumeSpecName "kube-api-access-ljwmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.871514 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config" (OuterVolumeSpecName: "config") pod "9383f078-40c5-48ab-937b-31593364ecc7" (UID: "9383f078-40c5-48ab-937b-31593364ecc7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.871803 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9383f078-40c5-48ab-937b-31593364ecc7" (UID: "9383f078-40c5-48ab-937b-31593364ecc7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.901711 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.901743 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljwmr\" (UniqueName: \"kubernetes.io/projected/9383f078-40c5-48ab-937b-31593364ecc7-kube-api-access-ljwmr\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:48 crc kubenswrapper[5018]: I1014 08:41:48.901754 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9383f078-40c5-48ab-937b-31593364ecc7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.345019 5018 generic.go:334] "Generic (PLEG): container finished" podID="9383f078-40c5-48ab-937b-31593364ecc7" containerID="578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde" exitCode=0 Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.345100 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" event={"ID":"9383f078-40c5-48ab-937b-31593364ecc7","Type":"ContainerDied","Data":"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde"} Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.345214 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" event={"ID":"9383f078-40c5-48ab-937b-31593364ecc7","Type":"ContainerDied","Data":"333bca4329bdd793ae9b73e23e3b7dec13a9034a2c971d43dc1196a7bd03f63c"} Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.345251 5018 scope.go:117] "RemoveContainer" containerID="578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.345116 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-kwvdx" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.438189 5018 scope.go:117] "RemoveContainer" containerID="e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.446674 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.457195 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-kwvdx"] Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.472894 5018 scope.go:117] "RemoveContainer" containerID="578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde" Oct 14 08:41:49 crc kubenswrapper[5018]: E1014 08:41:49.474057 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde\": container with ID starting with 578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde not found: ID does not exist" containerID="578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.474095 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde"} err="failed to get container status \"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde\": rpc error: code = NotFound desc = could not find container \"578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde\": container with ID starting with 578e5d8274ffbf1e21599d863017e2c8d866b913f22a4e94acef18abdd261bde not found: ID does not exist" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.474119 5018 scope.go:117] "RemoveContainer" containerID="e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d" Oct 14 08:41:49 crc kubenswrapper[5018]: E1014 08:41:49.474826 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d\": container with ID starting with e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d not found: ID does not exist" containerID="e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d" Oct 14 08:41:49 crc kubenswrapper[5018]: I1014 08:41:49.474903 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d"} err="failed to get container status \"e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d\": rpc error: code = NotFound desc = could not find container \"e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d\": container with ID starting with e243c82040ba7a9cc5f1b8681526cb3248fad8eadedc92c58e663566b7fd9d4d not found: ID does not exist" Oct 14 08:41:50 crc kubenswrapper[5018]: I1014 08:41:50.622399 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9383f078-40c5-48ab-937b-31593364ecc7" path="/var/lib/kubelet/pods/9383f078-40c5-48ab-937b-31593364ecc7/volumes" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.020216 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:41:51 crc kubenswrapper[5018]: E1014 08:41:51.021069 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="dnsmasq-dns" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.021213 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="dnsmasq-dns" Oct 14 08:41:51 crc kubenswrapper[5018]: E1014 08:41:51.021370 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="init" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.021476 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="init" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.021902 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9383f078-40c5-48ab-937b-31593364ecc7" containerName="dnsmasq-dns" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.023350 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.029238 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.029481 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.029692 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pjh44" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.029863 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040228 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040282 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040311 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wqnq\" (UniqueName: \"kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040479 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040717 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040865 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.040982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.058174 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.142987 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143074 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143097 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143117 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wqnq\" (UniqueName: \"kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143155 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143193 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.143229 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.144756 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.145025 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.145354 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.148310 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.150337 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.153571 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.164104 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wqnq\" (UniqueName: \"kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq\") pod \"ovn-northd-0\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.357245 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.888185 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:41:51 crc kubenswrapper[5018]: W1014 08:41:51.905562 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d8e88d0_a4e0_4206_bce3_998742afbdc5.slice/crio-85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958 WatchSource:0}: Error finding container 85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958: Status 404 returned error can't find the container with id 85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958 Oct 14 08:41:51 crc kubenswrapper[5018]: I1014 08:41:51.922348 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:41:52 crc kubenswrapper[5018]: I1014 08:41:52.382255 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerStarted","Data":"85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958"} Oct 14 08:41:54 crc kubenswrapper[5018]: I1014 08:41:54.409774 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerStarted","Data":"0cf0d1c6030e09c2eda77accb9e964aeb0ff1f0dbf949d67d6d6824cf055ea0a"} Oct 14 08:41:54 crc kubenswrapper[5018]: I1014 08:41:54.410401 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerStarted","Data":"d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b"} Oct 14 08:41:54 crc kubenswrapper[5018]: I1014 08:41:54.410432 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 08:41:54 crc kubenswrapper[5018]: I1014 08:41:54.439834 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.137980282 podStartE2EDuration="4.439814094s" podCreationTimestamp="2025-10-14 08:41:50 +0000 UTC" firstStartedPulling="2025-10-14 08:41:51.922056647 +0000 UTC m=+6728.506103274" lastFinishedPulling="2025-10-14 08:41:53.223890429 +0000 UTC m=+6729.807937086" observedRunningTime="2025-10-14 08:41:54.437031025 +0000 UTC m=+6731.021077692" watchObservedRunningTime="2025-10-14 08:41:54.439814094 +0000 UTC m=+6731.023860731" Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.654296 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-p45q9"] Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.656488 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-p45q9" Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.661349 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-p45q9"] Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.848543 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6mg9\" (UniqueName: \"kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9\") pod \"keystone-db-create-p45q9\" (UID: \"c76f3114-2b8a-4dab-a099-7934a05b33aa\") " pod="openstack/keystone-db-create-p45q9" Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.950533 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6mg9\" (UniqueName: \"kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9\") pod \"keystone-db-create-p45q9\" (UID: \"c76f3114-2b8a-4dab-a099-7934a05b33aa\") " pod="openstack/keystone-db-create-p45q9" Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.967788 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6mg9\" (UniqueName: \"kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9\") pod \"keystone-db-create-p45q9\" (UID: \"c76f3114-2b8a-4dab-a099-7934a05b33aa\") " pod="openstack/keystone-db-create-p45q9" Oct 14 08:41:59 crc kubenswrapper[5018]: I1014 08:41:59.985508 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-p45q9" Oct 14 08:42:00 crc kubenswrapper[5018]: I1014 08:42:00.475270 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-p45q9"] Oct 14 08:42:01 crc kubenswrapper[5018]: I1014 08:42:01.491545 5018 generic.go:334] "Generic (PLEG): container finished" podID="c76f3114-2b8a-4dab-a099-7934a05b33aa" containerID="17158a8fa8c3a15c6b14e71d5b226e53334837e0ea5686edc05334d9ee3734c6" exitCode=0 Oct 14 08:42:01 crc kubenswrapper[5018]: I1014 08:42:01.491694 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-p45q9" event={"ID":"c76f3114-2b8a-4dab-a099-7934a05b33aa","Type":"ContainerDied","Data":"17158a8fa8c3a15c6b14e71d5b226e53334837e0ea5686edc05334d9ee3734c6"} Oct 14 08:42:01 crc kubenswrapper[5018]: I1014 08:42:01.492013 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-p45q9" event={"ID":"c76f3114-2b8a-4dab-a099-7934a05b33aa","Type":"ContainerStarted","Data":"e37f26821fd7bad8f5c605fcdb2c274f8c53ad99fb152333cdec907cffe94c2f"} Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.021078 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-p45q9" Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.218513 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6mg9\" (UniqueName: \"kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9\") pod \"c76f3114-2b8a-4dab-a099-7934a05b33aa\" (UID: \"c76f3114-2b8a-4dab-a099-7934a05b33aa\") " Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.230145 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9" (OuterVolumeSpecName: "kube-api-access-q6mg9") pod "c76f3114-2b8a-4dab-a099-7934a05b33aa" (UID: "c76f3114-2b8a-4dab-a099-7934a05b33aa"). InnerVolumeSpecName "kube-api-access-q6mg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.320808 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6mg9\" (UniqueName: \"kubernetes.io/projected/c76f3114-2b8a-4dab-a099-7934a05b33aa-kube-api-access-q6mg9\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.517983 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-p45q9" event={"ID":"c76f3114-2b8a-4dab-a099-7934a05b33aa","Type":"ContainerDied","Data":"e37f26821fd7bad8f5c605fcdb2c274f8c53ad99fb152333cdec907cffe94c2f"} Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.518039 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e37f26821fd7bad8f5c605fcdb2c274f8c53ad99fb152333cdec907cffe94c2f" Oct 14 08:42:03 crc kubenswrapper[5018]: I1014 08:42:03.518119 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-p45q9" Oct 14 08:42:06 crc kubenswrapper[5018]: I1014 08:42:06.447449 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.647879 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-919f-account-create-vdsxq"] Oct 14 08:42:09 crc kubenswrapper[5018]: E1014 08:42:09.648478 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76f3114-2b8a-4dab-a099-7934a05b33aa" containerName="mariadb-database-create" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.648491 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76f3114-2b8a-4dab-a099-7934a05b33aa" containerName="mariadb-database-create" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.648669 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c76f3114-2b8a-4dab-a099-7934a05b33aa" containerName="mariadb-database-create" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.649226 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.650925 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.659205 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-919f-account-create-vdsxq"] Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.750380 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvs2\" (UniqueName: \"kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2\") pod \"keystone-919f-account-create-vdsxq\" (UID: \"5e155fce-f943-4909-a794-3d8c3d2bb036\") " pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.852299 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvs2\" (UniqueName: \"kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2\") pod \"keystone-919f-account-create-vdsxq\" (UID: \"5e155fce-f943-4909-a794-3d8c3d2bb036\") " pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.885401 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvs2\" (UniqueName: \"kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2\") pod \"keystone-919f-account-create-vdsxq\" (UID: \"5e155fce-f943-4909-a794-3d8c3d2bb036\") " pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:09 crc kubenswrapper[5018]: I1014 08:42:09.969698 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:10 crc kubenswrapper[5018]: I1014 08:42:10.520596 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-919f-account-create-vdsxq"] Oct 14 08:42:10 crc kubenswrapper[5018]: W1014 08:42:10.529902 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e155fce_f943_4909_a794_3d8c3d2bb036.slice/crio-7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b WatchSource:0}: Error finding container 7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b: Status 404 returned error can't find the container with id 7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b Oct 14 08:42:10 crc kubenswrapper[5018]: I1014 08:42:10.601851 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-919f-account-create-vdsxq" event={"ID":"5e155fce-f943-4909-a794-3d8c3d2bb036","Type":"ContainerStarted","Data":"7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b"} Oct 14 08:42:11 crc kubenswrapper[5018]: I1014 08:42:11.611530 5018 generic.go:334] "Generic (PLEG): container finished" podID="5e155fce-f943-4909-a794-3d8c3d2bb036" containerID="a4f64eef94cfa3ab48e1dc2b370ab0b6f9daff4ffcbb1eceb12fadffc3f2ea61" exitCode=0 Oct 14 08:42:11 crc kubenswrapper[5018]: I1014 08:42:11.611581 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-919f-account-create-vdsxq" event={"ID":"5e155fce-f943-4909-a794-3d8c3d2bb036","Type":"ContainerDied","Data":"a4f64eef94cfa3ab48e1dc2b370ab0b6f9daff4ffcbb1eceb12fadffc3f2ea61"} Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.052271 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.207298 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmvs2\" (UniqueName: \"kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2\") pod \"5e155fce-f943-4909-a794-3d8c3d2bb036\" (UID: \"5e155fce-f943-4909-a794-3d8c3d2bb036\") " Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.218049 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2" (OuterVolumeSpecName: "kube-api-access-tmvs2") pod "5e155fce-f943-4909-a794-3d8c3d2bb036" (UID: "5e155fce-f943-4909-a794-3d8c3d2bb036"). InnerVolumeSpecName "kube-api-access-tmvs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.309832 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmvs2\" (UniqueName: \"kubernetes.io/projected/5e155fce-f943-4909-a794-3d8c3d2bb036-kube-api-access-tmvs2\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.630145 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-919f-account-create-vdsxq" event={"ID":"5e155fce-f943-4909-a794-3d8c3d2bb036","Type":"ContainerDied","Data":"7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b"} Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.630188 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7284f9fbcb0f1d4c1890d12239c8656bb0fc05ab468fd1c7bec7cd176d96250b" Oct 14 08:42:13 crc kubenswrapper[5018]: I1014 08:42:13.630217 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-919f-account-create-vdsxq" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.112739 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-v97s4"] Oct 14 08:42:15 crc kubenswrapper[5018]: E1014 08:42:15.113269 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e155fce-f943-4909-a794-3d8c3d2bb036" containerName="mariadb-account-create" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.113282 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e155fce-f943-4909-a794-3d8c3d2bb036" containerName="mariadb-account-create" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.113451 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e155fce-f943-4909-a794-3d8c3d2bb036" containerName="mariadb-account-create" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.113979 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.116404 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.116684 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xkzvc" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.117251 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.117496 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.134833 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-v97s4"] Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.241563 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.241908 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg698\" (UniqueName: \"kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.242130 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.344278 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.344453 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg698\" (UniqueName: \"kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.344510 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.351104 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.359328 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.367502 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg698\" (UniqueName: \"kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698\") pod \"keystone-db-sync-v97s4\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.435077 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:15 crc kubenswrapper[5018]: I1014 08:42:15.906715 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-v97s4"] Oct 14 08:42:15 crc kubenswrapper[5018]: W1014 08:42:15.917148 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded7661bd_4acd_4bc1_9095_528af96485a9.slice/crio-28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175 WatchSource:0}: Error finding container 28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175: Status 404 returned error can't find the container with id 28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175 Oct 14 08:42:16 crc kubenswrapper[5018]: I1014 08:42:16.659390 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v97s4" event={"ID":"ed7661bd-4acd-4bc1-9095-528af96485a9","Type":"ContainerStarted","Data":"28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175"} Oct 14 08:42:21 crc kubenswrapper[5018]: I1014 08:42:21.706612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v97s4" event={"ID":"ed7661bd-4acd-4bc1-9095-528af96485a9","Type":"ContainerStarted","Data":"24db27ccf8374e2c2a417d5c951352e914ac93536c05b816a358776448c6f0bb"} Oct 14 08:42:21 crc kubenswrapper[5018]: I1014 08:42:21.732518 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-v97s4" podStartSLOduration=1.850070229 podStartE2EDuration="6.732491002s" podCreationTimestamp="2025-10-14 08:42:15 +0000 UTC" firstStartedPulling="2025-10-14 08:42:15.919945958 +0000 UTC m=+6752.503992615" lastFinishedPulling="2025-10-14 08:42:20.802366751 +0000 UTC m=+6757.386413388" observedRunningTime="2025-10-14 08:42:21.730641469 +0000 UTC m=+6758.314688136" watchObservedRunningTime="2025-10-14 08:42:21.732491002 +0000 UTC m=+6758.316537669" Oct 14 08:42:22 crc kubenswrapper[5018]: I1014 08:42:22.717415 5018 generic.go:334] "Generic (PLEG): container finished" podID="ed7661bd-4acd-4bc1-9095-528af96485a9" containerID="24db27ccf8374e2c2a417d5c951352e914ac93536c05b816a358776448c6f0bb" exitCode=0 Oct 14 08:42:22 crc kubenswrapper[5018]: I1014 08:42:22.717493 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v97s4" event={"ID":"ed7661bd-4acd-4bc1-9095-528af96485a9","Type":"ContainerDied","Data":"24db27ccf8374e2c2a417d5c951352e914ac93536c05b816a358776448c6f0bb"} Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.193470 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.307946 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data\") pod \"ed7661bd-4acd-4bc1-9095-528af96485a9\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.308080 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle\") pod \"ed7661bd-4acd-4bc1-9095-528af96485a9\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.308178 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg698\" (UniqueName: \"kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698\") pod \"ed7661bd-4acd-4bc1-9095-528af96485a9\" (UID: \"ed7661bd-4acd-4bc1-9095-528af96485a9\") " Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.314944 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698" (OuterVolumeSpecName: "kube-api-access-kg698") pod "ed7661bd-4acd-4bc1-9095-528af96485a9" (UID: "ed7661bd-4acd-4bc1-9095-528af96485a9"). InnerVolumeSpecName "kube-api-access-kg698". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.349924 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed7661bd-4acd-4bc1-9095-528af96485a9" (UID: "ed7661bd-4acd-4bc1-9095-528af96485a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.383987 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data" (OuterVolumeSpecName: "config-data") pod "ed7661bd-4acd-4bc1-9095-528af96485a9" (UID: "ed7661bd-4acd-4bc1-9095-528af96485a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.410539 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.410926 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7661bd-4acd-4bc1-9095-528af96485a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.411222 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg698\" (UniqueName: \"kubernetes.io/projected/ed7661bd-4acd-4bc1-9095-528af96485a9-kube-api-access-kg698\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.742965 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v97s4" event={"ID":"ed7661bd-4acd-4bc1-9095-528af96485a9","Type":"ContainerDied","Data":"28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175"} Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.743071 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c9c6f0547519ae91f4a5d322a05711b50144ac44ffb980005032d2383e3175" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.743797 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v97s4" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.992881 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:42:24 crc kubenswrapper[5018]: E1014 08:42:24.993186 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7661bd-4acd-4bc1-9095-528af96485a9" containerName="keystone-db-sync" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.993201 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7661bd-4acd-4bc1-9095-528af96485a9" containerName="keystone-db-sync" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.993351 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7661bd-4acd-4bc1-9095-528af96485a9" containerName="keystone-db-sync" Oct 14 08:42:24 crc kubenswrapper[5018]: I1014 08:42:24.994125 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.016822 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.040757 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2bhcf"] Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.042077 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.044570 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.045021 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.049722 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xkzvc" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.050105 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.056117 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2bhcf"] Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.123983 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.124327 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.124634 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6chw\" (UniqueName: \"kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.124683 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.124728 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.226185 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.226781 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.226870 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227044 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6chw\" (UniqueName: \"kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227133 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227230 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227317 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkp48\" (UniqueName: \"kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227400 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227482 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227579 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227673 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.227055 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.228009 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.228322 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.228682 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.247872 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6chw\" (UniqueName: \"kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw\") pod \"dnsmasq-dns-746ddcd59f-h5s9f\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.310859 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.328960 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkp48\" (UniqueName: \"kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.329026 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.329068 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.329132 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.329251 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.329290 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.333081 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.333357 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.337063 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.338343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.338608 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.360116 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkp48\" (UniqueName: \"kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48\") pod \"keystone-bootstrap-2bhcf\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.645147 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.660079 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:25 crc kubenswrapper[5018]: I1014 08:42:25.754699 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" event={"ID":"742282bd-dd4d-4cdd-8233-1bd9b9744a36","Type":"ContainerStarted","Data":"ae875ee9f4569fd76d831863a64ed9aaa25d1953b20eded5e06275a499f65af0"} Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.125362 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2bhcf"] Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.768188 5018 generic.go:334] "Generic (PLEG): container finished" podID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerID="761be0a3631b10d732cdf168f2d855c2ef99e7a7ef0dedf25c189237a1783d35" exitCode=0 Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.768266 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" event={"ID":"742282bd-dd4d-4cdd-8233-1bd9b9744a36","Type":"ContainerDied","Data":"761be0a3631b10d732cdf168f2d855c2ef99e7a7ef0dedf25c189237a1783d35"} Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.771385 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bhcf" event={"ID":"c0eccbcc-e863-4129-94c5-312a7977567e","Type":"ContainerStarted","Data":"dcd4c395ebfeabaaa50fe723fd4991e8d8c4491d321f34937e11bdb35db53078"} Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.771435 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bhcf" event={"ID":"c0eccbcc-e863-4129-94c5-312a7977567e","Type":"ContainerStarted","Data":"5c26b5b6ba354686376a62166bced8ffff786cfe9d1bf837cc21bdd3929d4297"} Oct 14 08:42:26 crc kubenswrapper[5018]: I1014 08:42:26.853295 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2bhcf" podStartSLOduration=1.8532747760000001 podStartE2EDuration="1.853274776s" podCreationTimestamp="2025-10-14 08:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:42:26.840245605 +0000 UTC m=+6763.424292232" watchObservedRunningTime="2025-10-14 08:42:26.853274776 +0000 UTC m=+6763.437321403" Oct 14 08:42:27 crc kubenswrapper[5018]: I1014 08:42:27.784807 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" event={"ID":"742282bd-dd4d-4cdd-8233-1bd9b9744a36","Type":"ContainerStarted","Data":"331ce05842d5501a7cf8976dfe046373e456e44a33387cc341b5bd2004ebe517"} Oct 14 08:42:27 crc kubenswrapper[5018]: I1014 08:42:27.811951 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" podStartSLOduration=3.811927011 podStartE2EDuration="3.811927011s" podCreationTimestamp="2025-10-14 08:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:42:27.806524907 +0000 UTC m=+6764.390571544" watchObservedRunningTime="2025-10-14 08:42:27.811927011 +0000 UTC m=+6764.395973638" Oct 14 08:42:28 crc kubenswrapper[5018]: I1014 08:42:28.803770 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:29 crc kubenswrapper[5018]: I1014 08:42:29.816987 5018 generic.go:334] "Generic (PLEG): container finished" podID="c0eccbcc-e863-4129-94c5-312a7977567e" containerID="dcd4c395ebfeabaaa50fe723fd4991e8d8c4491d321f34937e11bdb35db53078" exitCode=0 Oct 14 08:42:29 crc kubenswrapper[5018]: I1014 08:42:29.817065 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bhcf" event={"ID":"c0eccbcc-e863-4129-94c5-312a7977567e","Type":"ContainerDied","Data":"dcd4c395ebfeabaaa50fe723fd4991e8d8c4491d321f34937e11bdb35db53078"} Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.281835 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449295 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkp48\" (UniqueName: \"kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449445 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449467 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449498 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449522 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.449545 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys\") pod \"c0eccbcc-e863-4129-94c5-312a7977567e\" (UID: \"c0eccbcc-e863-4129-94c5-312a7977567e\") " Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.458958 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts" (OuterVolumeSpecName: "scripts") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.458973 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.462845 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.463870 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48" (OuterVolumeSpecName: "kube-api-access-tkp48") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "kube-api-access-tkp48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.479752 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data" (OuterVolumeSpecName: "config-data") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.496951 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0eccbcc-e863-4129-94c5-312a7977567e" (UID: "c0eccbcc-e863-4129-94c5-312a7977567e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552303 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkp48\" (UniqueName: \"kubernetes.io/projected/c0eccbcc-e863-4129-94c5-312a7977567e-kube-api-access-tkp48\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552562 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552640 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552726 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552798 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.552876 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c0eccbcc-e863-4129-94c5-312a7977567e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.845982 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bhcf" event={"ID":"c0eccbcc-e863-4129-94c5-312a7977567e","Type":"ContainerDied","Data":"5c26b5b6ba354686376a62166bced8ffff786cfe9d1bf837cc21bdd3929d4297"} Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.846043 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c26b5b6ba354686376a62166bced8ffff786cfe9d1bf837cc21bdd3929d4297" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.846072 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bhcf" Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.961604 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2bhcf"] Oct 14 08:42:31 crc kubenswrapper[5018]: I1014 08:42:31.976684 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2bhcf"] Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.045750 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6wjb4"] Oct 14 08:42:32 crc kubenswrapper[5018]: E1014 08:42:32.046469 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0eccbcc-e863-4129-94c5-312a7977567e" containerName="keystone-bootstrap" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.046485 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0eccbcc-e863-4129-94c5-312a7977567e" containerName="keystone-bootstrap" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.046652 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0eccbcc-e863-4129-94c5-312a7977567e" containerName="keystone-bootstrap" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.047254 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.049634 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.049783 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.049871 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xkzvc" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.050155 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.058251 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6wjb4"] Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.163557 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.163753 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.163831 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6wb\" (UniqueName: \"kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.163973 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.164131 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.164209 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.266480 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.266960 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.267232 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.267415 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6wb\" (UniqueName: \"kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.267653 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.267851 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.274030 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.274298 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.275171 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.276379 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.279099 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.302300 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6wb\" (UniqueName: \"kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb\") pod \"keystone-bootstrap-6wjb4\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.367187 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.625391 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0eccbcc-e863-4129-94c5-312a7977567e" path="/var/lib/kubelet/pods/c0eccbcc-e863-4129-94c5-312a7977567e/volumes" Oct 14 08:42:32 crc kubenswrapper[5018]: I1014 08:42:32.929376 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6wjb4"] Oct 14 08:42:33 crc kubenswrapper[5018]: I1014 08:42:33.869896 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wjb4" event={"ID":"3ddd357a-32f5-4a4a-954a-50740f3d4e23","Type":"ContainerStarted","Data":"736a9779931d58dc29fa4c031e5b54d1b6eb2ad4fce588ed956b779339e83771"} Oct 14 08:42:33 crc kubenswrapper[5018]: I1014 08:42:33.871544 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wjb4" event={"ID":"3ddd357a-32f5-4a4a-954a-50740f3d4e23","Type":"ContainerStarted","Data":"96df50d374c866d9a2de8c356d70bb8cd429e692b323e4d6c4265ee89165d1ea"} Oct 14 08:42:33 crc kubenswrapper[5018]: I1014 08:42:33.907746 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6wjb4" podStartSLOduration=1.907685565 podStartE2EDuration="1.907685565s" podCreationTimestamp="2025-10-14 08:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:42:33.891125173 +0000 UTC m=+6770.475171830" watchObservedRunningTime="2025-10-14 08:42:33.907685565 +0000 UTC m=+6770.491732232" Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.312885 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.392963 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.393297 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="dnsmasq-dns" containerID="cri-o://6ec0fca037768af2c6ab54de01240b5c1ec59460aa82d49fee10b48b4e0cd173" gracePeriod=10 Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.889242 5018 generic.go:334] "Generic (PLEG): container finished" podID="3ddd357a-32f5-4a4a-954a-50740f3d4e23" containerID="736a9779931d58dc29fa4c031e5b54d1b6eb2ad4fce588ed956b779339e83771" exitCode=0 Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.889304 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wjb4" event={"ID":"3ddd357a-32f5-4a4a-954a-50740f3d4e23","Type":"ContainerDied","Data":"736a9779931d58dc29fa4c031e5b54d1b6eb2ad4fce588ed956b779339e83771"} Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.892437 5018 generic.go:334] "Generic (PLEG): container finished" podID="919e2736-0848-4179-a50f-23782208a439" containerID="6ec0fca037768af2c6ab54de01240b5c1ec59460aa82d49fee10b48b4e0cd173" exitCode=0 Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.892467 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" event={"ID":"919e2736-0848-4179-a50f-23782208a439","Type":"ContainerDied","Data":"6ec0fca037768af2c6ab54de01240b5c1ec59460aa82d49fee10b48b4e0cd173"} Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.892487 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" event={"ID":"919e2736-0848-4179-a50f-23782208a439","Type":"ContainerDied","Data":"339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649"} Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.892504 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="339839365aede87b6ace60f200b4e0d8daac9cad9312ede3cbf7855fd4d33649" Oct 14 08:42:35 crc kubenswrapper[5018]: I1014 08:42:35.931266 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.058700 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmq4k\" (UniqueName: \"kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k\") pod \"919e2736-0848-4179-a50f-23782208a439\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.058748 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb\") pod \"919e2736-0848-4179-a50f-23782208a439\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.058933 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc\") pod \"919e2736-0848-4179-a50f-23782208a439\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.058971 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config\") pod \"919e2736-0848-4179-a50f-23782208a439\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.059012 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb\") pod \"919e2736-0848-4179-a50f-23782208a439\" (UID: \"919e2736-0848-4179-a50f-23782208a439\") " Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.072884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k" (OuterVolumeSpecName: "kube-api-access-lmq4k") pod "919e2736-0848-4179-a50f-23782208a439" (UID: "919e2736-0848-4179-a50f-23782208a439"). InnerVolumeSpecName "kube-api-access-lmq4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.103384 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "919e2736-0848-4179-a50f-23782208a439" (UID: "919e2736-0848-4179-a50f-23782208a439"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.105224 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "919e2736-0848-4179-a50f-23782208a439" (UID: "919e2736-0848-4179-a50f-23782208a439"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.114041 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "919e2736-0848-4179-a50f-23782208a439" (UID: "919e2736-0848-4179-a50f-23782208a439"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.115331 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config" (OuterVolumeSpecName: "config") pod "919e2736-0848-4179-a50f-23782208a439" (UID: "919e2736-0848-4179-a50f-23782208a439"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.161165 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.161192 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.161203 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmq4k\" (UniqueName: \"kubernetes.io/projected/919e2736-0848-4179-a50f-23782208a439-kube-api-access-lmq4k\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.161212 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.161220 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/919e2736-0848-4179-a50f-23782208a439-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.902512 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc8959bc9-cn9t4" Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.954346 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:42:36 crc kubenswrapper[5018]: I1014 08:42:36.970232 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bc8959bc9-cn9t4"] Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.324664 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485418 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485481 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk6wb\" (UniqueName: \"kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485542 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485678 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485710 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.485772 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle\") pod \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\" (UID: \"3ddd357a-32f5-4a4a-954a-50740f3d4e23\") " Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.497334 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.498053 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.498458 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb" (OuterVolumeSpecName: "kube-api-access-bk6wb") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "kube-api-access-bk6wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.505614 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts" (OuterVolumeSpecName: "scripts") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.518810 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.522045 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data" (OuterVolumeSpecName: "config-data") pod "3ddd357a-32f5-4a4a-954a-50740f3d4e23" (UID: "3ddd357a-32f5-4a4a-954a-50740f3d4e23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.587979 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.588013 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.588025 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.588039 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.588051 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddd357a-32f5-4a4a-954a-50740f3d4e23-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.588063 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk6wb\" (UniqueName: \"kubernetes.io/projected/3ddd357a-32f5-4a4a-954a-50740f3d4e23-kube-api-access-bk6wb\") on node \"crc\" DevicePath \"\"" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.916865 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wjb4" event={"ID":"3ddd357a-32f5-4a4a-954a-50740f3d4e23","Type":"ContainerDied","Data":"96df50d374c866d9a2de8c356d70bb8cd429e692b323e4d6c4265ee89165d1ea"} Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.916957 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96df50d374c866d9a2de8c356d70bb8cd429e692b323e4d6c4265ee89165d1ea" Oct 14 08:42:37 crc kubenswrapper[5018]: I1014 08:42:37.916957 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wjb4" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.024719 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 08:42:38 crc kubenswrapper[5018]: E1014 08:42:38.025226 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="init" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.025261 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="init" Oct 14 08:42:38 crc kubenswrapper[5018]: E1014 08:42:38.025289 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddd357a-32f5-4a4a-954a-50740f3d4e23" containerName="keystone-bootstrap" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.025298 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddd357a-32f5-4a4a-954a-50740f3d4e23" containerName="keystone-bootstrap" Oct 14 08:42:38 crc kubenswrapper[5018]: E1014 08:42:38.025313 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="dnsmasq-dns" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.025324 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="dnsmasq-dns" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.025609 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="919e2736-0848-4179-a50f-23782208a439" containerName="dnsmasq-dns" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.025669 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddd357a-32f5-4a4a-954a-50740f3d4e23" containerName="keystone-bootstrap" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.026365 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.030453 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.030987 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.034090 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xkzvc" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.034090 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.035589 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.035696 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.044651 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.097935 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.097984 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098025 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098052 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098082 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098112 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098137 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.098156 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l45ft\" (UniqueName: \"kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199465 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199543 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199588 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199645 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199688 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199725 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199750 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l45ft\" (UniqueName: \"kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.199817 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.203653 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.203872 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.204068 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.204154 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.205045 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.208513 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.222387 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l45ft\" (UniqueName: \"kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.225867 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs\") pod \"keystone-5479495b9c-w7q9b\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.381598 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.613411 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919e2736-0848-4179-a50f-23782208a439" path="/var/lib/kubelet/pods/919e2736-0848-4179-a50f-23782208a439/volumes" Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.878271 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 08:42:38 crc kubenswrapper[5018]: I1014 08:42:38.929844 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5479495b9c-w7q9b" event={"ID":"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46","Type":"ContainerStarted","Data":"966129542e044847487d9d6857b9b661f73d0f0e9338a64e570c36e0b08f54c4"} Oct 14 08:42:39 crc kubenswrapper[5018]: I1014 08:42:39.940435 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5479495b9c-w7q9b" event={"ID":"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46","Type":"ContainerStarted","Data":"3b364f7b933d0c59e076d99185b783f5158d326517a316e15f35f3ba0599def5"} Oct 14 08:42:39 crc kubenswrapper[5018]: I1014 08:42:39.940896 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:42:39 crc kubenswrapper[5018]: I1014 08:42:39.979262 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5479495b9c-w7q9b" podStartSLOduration=2.979236498 podStartE2EDuration="2.979236498s" podCreationTimestamp="2025-10-14 08:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:42:39.97402463 +0000 UTC m=+6776.558071297" watchObservedRunningTime="2025-10-14 08:42:39.979236498 +0000 UTC m=+6776.563283165" Oct 14 08:43:09 crc kubenswrapper[5018]: I1014 08:43:09.890136 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.720036 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.724675 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.731059 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-9ktcs" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.731116 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.736091 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.736476 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t6mp\" (UniqueName: \"kubernetes.io/projected/10c2603f-96bc-47c2-ae88-1935c2c8d786-kube-api-access-7t6mp\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.736571 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.736694 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.738332 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.740217 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.763769 5018 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10c2603f-96bc-47c2-ae88-1935c2c8d786\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T08:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T08:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T08:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T08:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:0468cb21803d466b2abfe00835cf1d2d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7t6mp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T08:43:14Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.768685 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: E1014 08:43:14.769946 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-7t6mp openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.781099 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.810558 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.812105 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.816450 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.826041 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.838306 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.838366 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74k5s\" (UniqueName: \"kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.838459 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.838493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t6mp\" (UniqueName: \"kubernetes.io/projected/10c2603f-96bc-47c2-ae88-1935c2c8d786-kube-api-access-7t6mp\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.838547 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.839693 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.839739 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.839769 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: E1014 08:43:14.841115 5018 projected.go:194] Error preparing data for projected volume kube-api-access-7t6mp for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (10c2603f-96bc-47c2-ae88-1935c2c8d786) does not match the UID in record. The object might have been deleted and then recreated Oct 14 08:43:14 crc kubenswrapper[5018]: E1014 08:43:14.841186 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/10c2603f-96bc-47c2-ae88-1935c2c8d786-kube-api-access-7t6mp podName:10c2603f-96bc-47c2-ae88-1935c2c8d786 nodeName:}" failed. No retries permitted until 2025-10-14 08:43:15.341168114 +0000 UTC m=+6811.925214741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-7t6mp" (UniqueName: "kubernetes.io/projected/10c2603f-96bc-47c2-ae88-1935c2c8d786-kube-api-access-7t6mp") pod "openstackclient" (UID: "10c2603f-96bc-47c2-ae88-1935c2c8d786") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (10c2603f-96bc-47c2-ae88-1935c2c8d786) does not match the UID in record. The object might have been deleted and then recreated Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.841185 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.845574 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.846439 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle\") pod \"openstackclient\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.940894 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.940971 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74k5s\" (UniqueName: \"kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.941032 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.941075 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.942514 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.945069 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.945542 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:14 crc kubenswrapper[5018]: I1014 08:43:14.959161 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74k5s\" (UniqueName: \"kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s\") pod \"openstackclient\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " pod="openstack/openstackclient" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.131378 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.297419 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.300929 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.317039 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.320806 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.348104 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret\") pod \"10c2603f-96bc-47c2-ae88-1935c2c8d786\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.348163 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle\") pod \"10c2603f-96bc-47c2-ae88-1935c2c8d786\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.348208 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config\") pod \"10c2603f-96bc-47c2-ae88-1935c2c8d786\" (UID: \"10c2603f-96bc-47c2-ae88-1935c2c8d786\") " Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.348567 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t6mp\" (UniqueName: \"kubernetes.io/projected/10c2603f-96bc-47c2-ae88-1935c2c8d786-kube-api-access-7t6mp\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.348884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "10c2603f-96bc-47c2-ae88-1935c2c8d786" (UID: "10c2603f-96bc-47c2-ae88-1935c2c8d786"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.353179 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10c2603f-96bc-47c2-ae88-1935c2c8d786" (UID: "10c2603f-96bc-47c2-ae88-1935c2c8d786"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.359838 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "10c2603f-96bc-47c2-ae88-1935c2c8d786" (UID: "10c2603f-96bc-47c2-ae88-1935c2c8d786"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.450090 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.450125 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2603f-96bc-47c2-ae88-1935c2c8d786-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.450138 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/10c2603f-96bc-47c2-ae88-1935c2c8d786-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:15 crc kubenswrapper[5018]: I1014 08:43:15.676854 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:43:16 crc kubenswrapper[5018]: I1014 08:43:16.308033 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20c5d68f-66eb-4918-85fc-ce1c18960eb7","Type":"ContainerStarted","Data":"ac0dd6ee14ece45ca7cd80206f8a2f85541bd56a49c59c32eb64b55847daba88"} Oct 14 08:43:16 crc kubenswrapper[5018]: I1014 08:43:16.308066 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:43:16 crc kubenswrapper[5018]: I1014 08:43:16.311224 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" Oct 14 08:43:16 crc kubenswrapper[5018]: I1014 08:43:16.326422 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" Oct 14 08:43:16 crc kubenswrapper[5018]: I1014 08:43:16.614161 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c2603f-96bc-47c2-ae88-1935c2c8d786" path="/var/lib/kubelet/pods/10c2603f-96bc-47c2-ae88-1935c2c8d786/volumes" Oct 14 08:43:26 crc kubenswrapper[5018]: I1014 08:43:26.397708 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20c5d68f-66eb-4918-85fc-ce1c18960eb7","Type":"ContainerStarted","Data":"15a85a608abcc9b92abd5e6818da1d44323d70efb24d37f1fdb4f4fd60ac9b4c"} Oct 14 08:43:26 crc kubenswrapper[5018]: I1014 08:43:26.425363 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.058100918 podStartE2EDuration="12.425346478s" podCreationTimestamp="2025-10-14 08:43:14 +0000 UTC" firstStartedPulling="2025-10-14 08:43:15.682670951 +0000 UTC m=+6812.266717578" lastFinishedPulling="2025-10-14 08:43:26.049916481 +0000 UTC m=+6822.633963138" observedRunningTime="2025-10-14 08:43:26.421498728 +0000 UTC m=+6823.005545355" watchObservedRunningTime="2025-10-14 08:43:26.425346478 +0000 UTC m=+6823.009393105" Oct 14 08:43:32 crc kubenswrapper[5018]: I1014 08:43:32.463749 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:43:32 crc kubenswrapper[5018]: I1014 08:43:32.464428 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:44:02 crc kubenswrapper[5018]: I1014 08:44:02.463615 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:44:02 crc kubenswrapper[5018]: I1014 08:44:02.464522 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.063546 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.067814 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.083333 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.187678 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.187775 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpdqq\" (UniqueName: \"kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.187956 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.290021 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.290256 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.290336 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpdqq\" (UniqueName: \"kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.290668 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.290732 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.315379 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpdqq\" (UniqueName: \"kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq\") pod \"community-operators-szzhx\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.433921 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:16 crc kubenswrapper[5018]: I1014 08:44:16.953903 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:16 crc kubenswrapper[5018]: W1014 08:44:16.965408 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0398b73_5bae_4f77_a915_f84763ae8594.slice/crio-7d2f4bbdeb6abb6a8b614d6f6c961b3fe2c8e1d8ec0599f2de1bf632be2f9f7e WatchSource:0}: Error finding container 7d2f4bbdeb6abb6a8b614d6f6c961b3fe2c8e1d8ec0599f2de1bf632be2f9f7e: Status 404 returned error can't find the container with id 7d2f4bbdeb6abb6a8b614d6f6c961b3fe2c8e1d8ec0599f2de1bf632be2f9f7e Oct 14 08:44:17 crc kubenswrapper[5018]: I1014 08:44:17.981610 5018 generic.go:334] "Generic (PLEG): container finished" podID="d0398b73-5bae-4f77-a915-f84763ae8594" containerID="87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9" exitCode=0 Oct 14 08:44:17 crc kubenswrapper[5018]: I1014 08:44:17.982022 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerDied","Data":"87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9"} Oct 14 08:44:17 crc kubenswrapper[5018]: I1014 08:44:17.982227 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerStarted","Data":"7d2f4bbdeb6abb6a8b614d6f6c961b3fe2c8e1d8ec0599f2de1bf632be2f9f7e"} Oct 14 08:44:18 crc kubenswrapper[5018]: I1014 08:44:18.994291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerStarted","Data":"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7"} Oct 14 08:44:20 crc kubenswrapper[5018]: I1014 08:44:20.003969 5018 generic.go:334] "Generic (PLEG): container finished" podID="d0398b73-5bae-4f77-a915-f84763ae8594" containerID="69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7" exitCode=0 Oct 14 08:44:20 crc kubenswrapper[5018]: I1014 08:44:20.004034 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerDied","Data":"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7"} Oct 14 08:44:22 crc kubenswrapper[5018]: I1014 08:44:22.019353 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerStarted","Data":"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299"} Oct 14 08:44:22 crc kubenswrapper[5018]: I1014 08:44:22.040746 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-szzhx" podStartSLOduration=3.230493842 podStartE2EDuration="6.040722393s" podCreationTimestamp="2025-10-14 08:44:16 +0000 UTC" firstStartedPulling="2025-10-14 08:44:17.995318679 +0000 UTC m=+6874.579365336" lastFinishedPulling="2025-10-14 08:44:20.80554723 +0000 UTC m=+6877.389593887" observedRunningTime="2025-10-14 08:44:22.036016749 +0000 UTC m=+6878.620063386" watchObservedRunningTime="2025-10-14 08:44:22.040722393 +0000 UTC m=+6878.624769030" Oct 14 08:44:26 crc kubenswrapper[5018]: I1014 08:44:26.434447 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:26 crc kubenswrapper[5018]: I1014 08:44:26.434873 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:26 crc kubenswrapper[5018]: I1014 08:44:26.506483 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:27 crc kubenswrapper[5018]: I1014 08:44:27.139010 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:27 crc kubenswrapper[5018]: I1014 08:44:27.210950 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.099501 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-szzhx" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="registry-server" containerID="cri-o://ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299" gracePeriod=2 Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.655330 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.837890 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpdqq\" (UniqueName: \"kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq\") pod \"d0398b73-5bae-4f77-a915-f84763ae8594\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.838080 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities\") pod \"d0398b73-5bae-4f77-a915-f84763ae8594\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.838229 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content\") pod \"d0398b73-5bae-4f77-a915-f84763ae8594\" (UID: \"d0398b73-5bae-4f77-a915-f84763ae8594\") " Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.840907 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities" (OuterVolumeSpecName: "utilities") pod "d0398b73-5bae-4f77-a915-f84763ae8594" (UID: "d0398b73-5bae-4f77-a915-f84763ae8594"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.861284 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq" (OuterVolumeSpecName: "kube-api-access-hpdqq") pod "d0398b73-5bae-4f77-a915-f84763ae8594" (UID: "d0398b73-5bae-4f77-a915-f84763ae8594"). InnerVolumeSpecName "kube-api-access-hpdqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.940565 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0398b73-5bae-4f77-a915-f84763ae8594" (UID: "d0398b73-5bae-4f77-a915-f84763ae8594"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.940978 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.940996 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpdqq\" (UniqueName: \"kubernetes.io/projected/d0398b73-5bae-4f77-a915-f84763ae8594-kube-api-access-hpdqq\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:29 crc kubenswrapper[5018]: I1014 08:44:29.941009 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0398b73-5bae-4f77-a915-f84763ae8594-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.113157 5018 generic.go:334] "Generic (PLEG): container finished" podID="d0398b73-5bae-4f77-a915-f84763ae8594" containerID="ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299" exitCode=0 Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.113236 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerDied","Data":"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299"} Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.113297 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szzhx" event={"ID":"d0398b73-5bae-4f77-a915-f84763ae8594","Type":"ContainerDied","Data":"7d2f4bbdeb6abb6a8b614d6f6c961b3fe2c8e1d8ec0599f2de1bf632be2f9f7e"} Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.113309 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szzhx" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.113340 5018 scope.go:117] "RemoveContainer" containerID="ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.147152 5018 scope.go:117] "RemoveContainer" containerID="69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.193446 5018 scope.go:117] "RemoveContainer" containerID="87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.204162 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.205587 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-szzhx"] Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.255893 5018 scope.go:117] "RemoveContainer" containerID="ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299" Oct 14 08:44:30 crc kubenswrapper[5018]: E1014 08:44:30.256408 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299\": container with ID starting with ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299 not found: ID does not exist" containerID="ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.256453 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299"} err="failed to get container status \"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299\": rpc error: code = NotFound desc = could not find container \"ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299\": container with ID starting with ea102f322fefb6d66b1b4d243f751d69b4249b4735cdbe480bd611e6fd2ab299 not found: ID does not exist" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.256483 5018 scope.go:117] "RemoveContainer" containerID="69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7" Oct 14 08:44:30 crc kubenswrapper[5018]: E1014 08:44:30.256925 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7\": container with ID starting with 69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7 not found: ID does not exist" containerID="69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.256984 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7"} err="failed to get container status \"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7\": rpc error: code = NotFound desc = could not find container \"69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7\": container with ID starting with 69463df99c8c3e64f4489ee473423496d3e7b57322f10a72e287f0676574f6a7 not found: ID does not exist" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.257026 5018 scope.go:117] "RemoveContainer" containerID="87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9" Oct 14 08:44:30 crc kubenswrapper[5018]: E1014 08:44:30.257376 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9\": container with ID starting with 87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9 not found: ID does not exist" containerID="87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.257418 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9"} err="failed to get container status \"87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9\": rpc error: code = NotFound desc = could not find container \"87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9\": container with ID starting with 87de1643a932baea6ffd660e4116497afdf9e9f65c5570f565e1578c2d3b17d9 not found: ID does not exist" Oct 14 08:44:30 crc kubenswrapper[5018]: I1014 08:44:30.623996 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" path="/var/lib/kubelet/pods/d0398b73-5bae-4f77-a915-f84763ae8594/volumes" Oct 14 08:44:32 crc kubenswrapper[5018]: I1014 08:44:32.464263 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:44:32 crc kubenswrapper[5018]: I1014 08:44:32.464892 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:44:32 crc kubenswrapper[5018]: I1014 08:44:32.464969 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:44:32 crc kubenswrapper[5018]: I1014 08:44:32.466047 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:44:32 crc kubenswrapper[5018]: I1014 08:44:32.466147 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee" gracePeriod=600 Oct 14 08:44:33 crc kubenswrapper[5018]: I1014 08:44:33.158084 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee" exitCode=0 Oct 14 08:44:33 crc kubenswrapper[5018]: I1014 08:44:33.158176 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee"} Oct 14 08:44:33 crc kubenswrapper[5018]: I1014 08:44:33.158810 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44"} Oct 14 08:44:33 crc kubenswrapper[5018]: I1014 08:44:33.158845 5018 scope.go:117] "RemoveContainer" containerID="4ec057b6ab29601d3d2370a86c7959a77d1bb96342e3bdfe3b7bcacb3319934c" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.111816 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-htt58"] Oct 14 08:44:47 crc kubenswrapper[5018]: E1014 08:44:47.112788 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="extract-content" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.112803 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="extract-content" Oct 14 08:44:47 crc kubenswrapper[5018]: E1014 08:44:47.112835 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="registry-server" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.112844 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="registry-server" Oct 14 08:44:47 crc kubenswrapper[5018]: E1014 08:44:47.112857 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="extract-utilities" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.112866 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="extract-utilities" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.113097 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0398b73-5bae-4f77-a915-f84763ae8594" containerName="registry-server" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.113786 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htt58" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.120445 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htt58"] Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.227982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4l9x\" (UniqueName: \"kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x\") pod \"barbican-db-create-htt58\" (UID: \"cf2f6095-4fe6-43fa-a482-5767e720f595\") " pod="openstack/barbican-db-create-htt58" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.329817 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4l9x\" (UniqueName: \"kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x\") pod \"barbican-db-create-htt58\" (UID: \"cf2f6095-4fe6-43fa-a482-5767e720f595\") " pod="openstack/barbican-db-create-htt58" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.360646 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4l9x\" (UniqueName: \"kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x\") pod \"barbican-db-create-htt58\" (UID: \"cf2f6095-4fe6-43fa-a482-5767e720f595\") " pod="openstack/barbican-db-create-htt58" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.439226 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htt58" Oct 14 08:44:47 crc kubenswrapper[5018]: I1014 08:44:47.982518 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htt58"] Oct 14 08:44:47 crc kubenswrapper[5018]: W1014 08:44:47.984582 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf2f6095_4fe6_43fa_a482_5767e720f595.slice/crio-30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17 WatchSource:0}: Error finding container 30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17: Status 404 returned error can't find the container with id 30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17 Oct 14 08:44:48 crc kubenswrapper[5018]: I1014 08:44:48.318501 5018 generic.go:334] "Generic (PLEG): container finished" podID="cf2f6095-4fe6-43fa-a482-5767e720f595" containerID="ed974f0484876dd1fa4b142f26956a40a8b3537cf9521a9601f07815d7625c75" exitCode=0 Oct 14 08:44:48 crc kubenswrapper[5018]: I1014 08:44:48.318569 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htt58" event={"ID":"cf2f6095-4fe6-43fa-a482-5767e720f595","Type":"ContainerDied","Data":"ed974f0484876dd1fa4b142f26956a40a8b3537cf9521a9601f07815d7625c75"} Oct 14 08:44:48 crc kubenswrapper[5018]: I1014 08:44:48.318929 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htt58" event={"ID":"cf2f6095-4fe6-43fa-a482-5767e720f595","Type":"ContainerStarted","Data":"30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17"} Oct 14 08:44:49 crc kubenswrapper[5018]: I1014 08:44:49.757117 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htt58" Oct 14 08:44:49 crc kubenswrapper[5018]: I1014 08:44:49.877270 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4l9x\" (UniqueName: \"kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x\") pod \"cf2f6095-4fe6-43fa-a482-5767e720f595\" (UID: \"cf2f6095-4fe6-43fa-a482-5767e720f595\") " Oct 14 08:44:49 crc kubenswrapper[5018]: I1014 08:44:49.886879 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x" (OuterVolumeSpecName: "kube-api-access-h4l9x") pod "cf2f6095-4fe6-43fa-a482-5767e720f595" (UID: "cf2f6095-4fe6-43fa-a482-5767e720f595"). InnerVolumeSpecName "kube-api-access-h4l9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:49 crc kubenswrapper[5018]: I1014 08:44:49.979114 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4l9x\" (UniqueName: \"kubernetes.io/projected/cf2f6095-4fe6-43fa-a482-5767e720f595-kube-api-access-h4l9x\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:50 crc kubenswrapper[5018]: I1014 08:44:50.341463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htt58" event={"ID":"cf2f6095-4fe6-43fa-a482-5767e720f595","Type":"ContainerDied","Data":"30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17"} Oct 14 08:44:50 crc kubenswrapper[5018]: I1014 08:44:50.341505 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30744d4b0db0806c2287726f838cfe65a7e13dd0ad89127a9b47df97d27e0b17" Oct 14 08:44:50 crc kubenswrapper[5018]: I1014 08:44:50.341565 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htt58" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.124745 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3632-account-create-94rwz"] Oct 14 08:44:57 crc kubenswrapper[5018]: E1014 08:44:57.126314 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf2f6095-4fe6-43fa-a482-5767e720f595" containerName="mariadb-database-create" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.126335 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf2f6095-4fe6-43fa-a482-5767e720f595" containerName="mariadb-database-create" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.126603 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf2f6095-4fe6-43fa-a482-5767e720f595" containerName="mariadb-database-create" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.127721 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.132328 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.137990 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3632-account-create-94rwz"] Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.226985 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdlqz\" (UniqueName: \"kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz\") pod \"barbican-3632-account-create-94rwz\" (UID: \"ee9a074e-cd7c-4cb9-bde2-fe324ab51708\") " pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.329363 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdlqz\" (UniqueName: \"kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz\") pod \"barbican-3632-account-create-94rwz\" (UID: \"ee9a074e-cd7c-4cb9-bde2-fe324ab51708\") " pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.358062 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdlqz\" (UniqueName: \"kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz\") pod \"barbican-3632-account-create-94rwz\" (UID: \"ee9a074e-cd7c-4cb9-bde2-fe324ab51708\") " pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.470180 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:57 crc kubenswrapper[5018]: I1014 08:44:57.979695 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3632-account-create-94rwz"] Oct 14 08:44:58 crc kubenswrapper[5018]: I1014 08:44:58.443603 5018 generic.go:334] "Generic (PLEG): container finished" podID="ee9a074e-cd7c-4cb9-bde2-fe324ab51708" containerID="79d84bb2589c90bf3af60e2e4c20a9e0c372807e8dd8a0afc53500cc2f00f05b" exitCode=0 Oct 14 08:44:58 crc kubenswrapper[5018]: I1014 08:44:58.443707 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3632-account-create-94rwz" event={"ID":"ee9a074e-cd7c-4cb9-bde2-fe324ab51708","Type":"ContainerDied","Data":"79d84bb2589c90bf3af60e2e4c20a9e0c372807e8dd8a0afc53500cc2f00f05b"} Oct 14 08:44:58 crc kubenswrapper[5018]: I1014 08:44:58.443745 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3632-account-create-94rwz" event={"ID":"ee9a074e-cd7c-4cb9-bde2-fe324ab51708","Type":"ContainerStarted","Data":"34485d21bb0db0be66c36b53a04e32527a1b9504da893f09aaf68e0d6b729229"} Oct 14 08:44:59 crc kubenswrapper[5018]: I1014 08:44:59.792633 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:44:59 crc kubenswrapper[5018]: I1014 08:44:59.884259 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdlqz\" (UniqueName: \"kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz\") pod \"ee9a074e-cd7c-4cb9-bde2-fe324ab51708\" (UID: \"ee9a074e-cd7c-4cb9-bde2-fe324ab51708\") " Oct 14 08:44:59 crc kubenswrapper[5018]: I1014 08:44:59.896357 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz" (OuterVolumeSpecName: "kube-api-access-vdlqz") pod "ee9a074e-cd7c-4cb9-bde2-fe324ab51708" (UID: "ee9a074e-cd7c-4cb9-bde2-fe324ab51708"). InnerVolumeSpecName "kube-api-access-vdlqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:59 crc kubenswrapper[5018]: I1014 08:44:59.986455 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdlqz\" (UniqueName: \"kubernetes.io/projected/ee9a074e-cd7c-4cb9-bde2-fe324ab51708-kube-api-access-vdlqz\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.166202 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2"] Oct 14 08:45:00 crc kubenswrapper[5018]: E1014 08:45:00.166935 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9a074e-cd7c-4cb9-bde2-fe324ab51708" containerName="mariadb-account-create" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.167072 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9a074e-cd7c-4cb9-bde2-fe324ab51708" containerName="mariadb-account-create" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.167483 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9a074e-cd7c-4cb9-bde2-fe324ab51708" containerName="mariadb-account-create" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.168471 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.171787 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.171799 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.175085 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2"] Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.291039 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.291110 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.291186 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7fm4\" (UniqueName: \"kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.392938 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7fm4\" (UniqueName: \"kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.393101 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.393142 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.394842 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.397524 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.418839 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7fm4\" (UniqueName: \"kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4\") pod \"collect-profiles-29340525-wjfk2\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.463186 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3632-account-create-94rwz" event={"ID":"ee9a074e-cd7c-4cb9-bde2-fe324ab51708","Type":"ContainerDied","Data":"34485d21bb0db0be66c36b53a04e32527a1b9504da893f09aaf68e0d6b729229"} Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.463233 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34485d21bb0db0be66c36b53a04e32527a1b9504da893f09aaf68e0d6b729229" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.463252 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3632-account-create-94rwz" Oct 14 08:45:00 crc kubenswrapper[5018]: I1014 08:45:00.541017 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:01 crc kubenswrapper[5018]: I1014 08:45:00.858788 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2"] Oct 14 08:45:01 crc kubenswrapper[5018]: W1014 08:45:00.871907 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a00fb58_e967_4ffa_8400_c68cdba8739d.slice/crio-16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c WatchSource:0}: Error finding container 16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c: Status 404 returned error can't find the container with id 16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c Oct 14 08:45:01 crc kubenswrapper[5018]: I1014 08:45:01.471565 5018 generic.go:334] "Generic (PLEG): container finished" podID="6a00fb58-e967-4ffa-8400-c68cdba8739d" containerID="932bf663b1d1c74ce9c75310cd6fe020eb06271de0a1907e2da276bc13cd3a1e" exitCode=0 Oct 14 08:45:01 crc kubenswrapper[5018]: I1014 08:45:01.471743 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" event={"ID":"6a00fb58-e967-4ffa-8400-c68cdba8739d","Type":"ContainerDied","Data":"932bf663b1d1c74ce9c75310cd6fe020eb06271de0a1907e2da276bc13cd3a1e"} Oct 14 08:45:01 crc kubenswrapper[5018]: I1014 08:45:01.471883 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" event={"ID":"6a00fb58-e967-4ffa-8400-c68cdba8739d","Type":"ContainerStarted","Data":"16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c"} Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.341830 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-z5s89"] Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.343206 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.345745 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.347898 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tljlw" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.363195 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z5s89"] Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.428646 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tccbx\" (UniqueName: \"kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.428727 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.428836 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.529928 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.530470 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tccbx\" (UniqueName: \"kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.530647 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.562419 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.563144 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.567277 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tccbx\" (UniqueName: \"kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx\") pod \"barbican-db-sync-z5s89\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.680014 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:02 crc kubenswrapper[5018]: I1014 08:45:02.894042 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.040955 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume\") pod \"6a00fb58-e967-4ffa-8400-c68cdba8739d\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.041127 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7fm4\" (UniqueName: \"kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4\") pod \"6a00fb58-e967-4ffa-8400-c68cdba8739d\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.041241 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume\") pod \"6a00fb58-e967-4ffa-8400-c68cdba8739d\" (UID: \"6a00fb58-e967-4ffa-8400-c68cdba8739d\") " Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.041741 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume" (OuterVolumeSpecName: "config-volume") pod "6a00fb58-e967-4ffa-8400-c68cdba8739d" (UID: "6a00fb58-e967-4ffa-8400-c68cdba8739d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.046348 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6a00fb58-e967-4ffa-8400-c68cdba8739d" (UID: "6a00fb58-e967-4ffa-8400-c68cdba8739d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.047556 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4" (OuterVolumeSpecName: "kube-api-access-j7fm4") pod "6a00fb58-e967-4ffa-8400-c68cdba8739d" (UID: "6a00fb58-e967-4ffa-8400-c68cdba8739d"). InnerVolumeSpecName "kube-api-access-j7fm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.130756 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z5s89"] Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.143276 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a00fb58-e967-4ffa-8400-c68cdba8739d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.143336 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a00fb58-e967-4ffa-8400-c68cdba8739d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.143357 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7fm4\" (UniqueName: \"kubernetes.io/projected/6a00fb58-e967-4ffa-8400-c68cdba8739d-kube-api-access-j7fm4\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.497959 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z5s89" event={"ID":"41b224e5-b8be-4cfa-9ec0-7796ae74e556","Type":"ContainerStarted","Data":"b44a68ab80b4fd49a19e0a1fb92d0ab60833a8ee37db70d1a62627880511e3b8"} Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.500425 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" event={"ID":"6a00fb58-e967-4ffa-8400-c68cdba8739d","Type":"ContainerDied","Data":"16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c"} Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.500461 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16869d4e030ac783e8641688463cbc9fde132dd884e72207d10271ad2bd0396c" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.500546 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-wjfk2" Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.984100 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6"] Oct 14 08:45:03 crc kubenswrapper[5018]: I1014 08:45:03.992581 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-962q6"] Oct 14 08:45:04 crc kubenswrapper[5018]: I1014 08:45:04.633316 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="688bb2d0-e6e4-495f-aa22-5dc6f890c7b8" path="/var/lib/kubelet/pods/688bb2d0-e6e4-495f-aa22-5dc6f890c7b8/volumes" Oct 14 08:45:07 crc kubenswrapper[5018]: I1014 08:45:07.239299 5018 scope.go:117] "RemoveContainer" containerID="045b55da320689ea87c32d82e821d8f765d4462d588ed2727b25b680b10f83c9" Oct 14 08:45:08 crc kubenswrapper[5018]: I1014 08:45:08.551882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z5s89" event={"ID":"41b224e5-b8be-4cfa-9ec0-7796ae74e556","Type":"ContainerStarted","Data":"b03117968d08499e531fc7721c3ce7ae44a008e802626016b03f23faa1b4b08f"} Oct 14 08:45:08 crc kubenswrapper[5018]: I1014 08:45:08.576760 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-z5s89" podStartSLOduration=1.438266055 podStartE2EDuration="6.576737774s" podCreationTimestamp="2025-10-14 08:45:02 +0000 UTC" firstStartedPulling="2025-10-14 08:45:03.132512993 +0000 UTC m=+6919.716559660" lastFinishedPulling="2025-10-14 08:45:08.270984732 +0000 UTC m=+6924.855031379" observedRunningTime="2025-10-14 08:45:08.566322347 +0000 UTC m=+6925.150369004" watchObservedRunningTime="2025-10-14 08:45:08.576737774 +0000 UTC m=+6925.160784411" Oct 14 08:45:10 crc kubenswrapper[5018]: I1014 08:45:10.575868 5018 generic.go:334] "Generic (PLEG): container finished" podID="41b224e5-b8be-4cfa-9ec0-7796ae74e556" containerID="b03117968d08499e531fc7721c3ce7ae44a008e802626016b03f23faa1b4b08f" exitCode=0 Oct 14 08:45:10 crc kubenswrapper[5018]: I1014 08:45:10.576348 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z5s89" event={"ID":"41b224e5-b8be-4cfa-9ec0-7796ae74e556","Type":"ContainerDied","Data":"b03117968d08499e531fc7721c3ce7ae44a008e802626016b03f23faa1b4b08f"} Oct 14 08:45:11 crc kubenswrapper[5018]: I1014 08:45:11.986648 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.113770 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tccbx\" (UniqueName: \"kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx\") pod \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.113962 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle\") pod \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.113993 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data\") pod \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\" (UID: \"41b224e5-b8be-4cfa-9ec0-7796ae74e556\") " Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.118748 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx" (OuterVolumeSpecName: "kube-api-access-tccbx") pod "41b224e5-b8be-4cfa-9ec0-7796ae74e556" (UID: "41b224e5-b8be-4cfa-9ec0-7796ae74e556"). InnerVolumeSpecName "kube-api-access-tccbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.119352 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "41b224e5-b8be-4cfa-9ec0-7796ae74e556" (UID: "41b224e5-b8be-4cfa-9ec0-7796ae74e556"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.142775 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41b224e5-b8be-4cfa-9ec0-7796ae74e556" (UID: "41b224e5-b8be-4cfa-9ec0-7796ae74e556"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.216772 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.216829 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41b224e5-b8be-4cfa-9ec0-7796ae74e556-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.216852 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tccbx\" (UniqueName: \"kubernetes.io/projected/41b224e5-b8be-4cfa-9ec0-7796ae74e556-kube-api-access-tccbx\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.598396 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z5s89" event={"ID":"41b224e5-b8be-4cfa-9ec0-7796ae74e556","Type":"ContainerDied","Data":"b44a68ab80b4fd49a19e0a1fb92d0ab60833a8ee37db70d1a62627880511e3b8"} Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.598453 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b44a68ab80b4fd49a19e0a1fb92d0ab60833a8ee37db70d1a62627880511e3b8" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.598498 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z5s89" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.891072 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:45:12 crc kubenswrapper[5018]: E1014 08:45:12.891358 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b224e5-b8be-4cfa-9ec0-7796ae74e556" containerName="barbican-db-sync" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.891374 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b224e5-b8be-4cfa-9ec0-7796ae74e556" containerName="barbican-db-sync" Oct 14 08:45:12 crc kubenswrapper[5018]: E1014 08:45:12.891411 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a00fb58-e967-4ffa-8400-c68cdba8739d" containerName="collect-profiles" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.891419 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a00fb58-e967-4ffa-8400-c68cdba8739d" containerName="collect-profiles" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.891560 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b224e5-b8be-4cfa-9ec0-7796ae74e556" containerName="barbican-db-sync" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.891581 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a00fb58-e967-4ffa-8400-c68cdba8739d" containerName="collect-profiles" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.892364 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.904575 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.923099 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.924936 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.928427 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.928657 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.930271 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tljlw" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.949126 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.950714 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.955910 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 08:45:12 crc kubenswrapper[5018]: I1014 08:45:12.978645 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.006576 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037036 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037088 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wck4\" (UniqueName: \"kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037232 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037276 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037312 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037339 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037365 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwwc8\" (UniqueName: \"kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037401 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037419 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037452 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt7w6\" (UniqueName: \"kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037468 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037579 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037599 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037634 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.037678 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.085740 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.087289 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.089472 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.107724 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140453 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwwc8\" (UniqueName: \"kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140508 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140531 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140557 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt7w6\" (UniqueName: \"kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140578 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140647 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140664 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140681 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140709 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140740 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140771 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wck4\" (UniqueName: \"kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140804 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140822 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.140860 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.149926 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.150397 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.151107 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.154033 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.162907 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.164116 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.165517 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.168263 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.170747 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt7w6\" (UniqueName: \"kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.173343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.176651 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.176911 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle\") pod \"barbican-keystone-listener-5dcf959744-4fkwc\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.181194 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwwc8\" (UniqueName: \"kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.181834 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wck4\" (UniqueName: \"kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4\") pod \"dnsmasq-dns-5977c8f7d7-b9c46\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.186168 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom\") pod \"barbican-worker-854795b57-c4dnm\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.209009 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.242464 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.242527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.242561 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnm9n\" (UniqueName: \"kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.242603 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.242702 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.260969 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.269983 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.344498 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.344539 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.344561 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.344592 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnm9n\" (UniqueName: \"kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.344644 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.347940 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.353390 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.353841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.355742 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.371883 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnm9n\" (UniqueName: \"kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n\") pod \"barbican-api-597c8fcb86-94j25\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.406457 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.840452 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 08:45:13 crc kubenswrapper[5018]: I1014 08:45:13.850605 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:45:14 crc kubenswrapper[5018]: W1014 08:45:14.000720 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b967f1d_3f1f_414f_87b9_c5e0ee989e23.slice/crio-1562d007f4b144548257c422a2911771c538a8901d3797f5334540d7bb00b957 WatchSource:0}: Error finding container 1562d007f4b144548257c422a2911771c538a8901d3797f5334540d7bb00b957: Status 404 returned error can't find the container with id 1562d007f4b144548257c422a2911771c538a8901d3797f5334540d7bb00b957 Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.001474 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.030926 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:14 crc kubenswrapper[5018]: W1014 08:45:14.046667 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a4c3de8_3125_47e2_8c69_bc71222ba3c3.slice/crio-5061ec448a7bb301146c1e3f2c07d08b38c6be9d52e5c8530876653a768e881a WatchSource:0}: Error finding container 5061ec448a7bb301146c1e3f2c07d08b38c6be9d52e5c8530876653a768e881a: Status 404 returned error can't find the container with id 5061ec448a7bb301146c1e3f2c07d08b38c6be9d52e5c8530876653a768e881a Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.634524 5018 generic.go:334] "Generic (PLEG): container finished" podID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerID="3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112" exitCode=0 Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.635809 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" event={"ID":"6275743b-3b10-4797-a5f3-ad613eb392f0","Type":"ContainerDied","Data":"3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112"} Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.635845 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" event={"ID":"6275743b-3b10-4797-a5f3-ad613eb392f0","Type":"ContainerStarted","Data":"01177809583a1274a9098bd68bffca2282ddbe7ebad8de5bb17eeb652e175a1d"} Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.637362 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerStarted","Data":"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708"} Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.637403 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerStarted","Data":"5061ec448a7bb301146c1e3f2c07d08b38c6be9d52e5c8530876653a768e881a"} Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.639779 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerStarted","Data":"90a6b0d8e9a7e5c57565c31b12754efe38f9de81afbd9e4c85903fa8a6636544"} Oct 14 08:45:14 crc kubenswrapper[5018]: I1014 08:45:14.640694 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerStarted","Data":"1562d007f4b144548257c422a2911771c538a8901d3797f5334540d7bb00b957"} Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.120944 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.122197 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.124248 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.125221 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.141309 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294577 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294654 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294703 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54qjw\" (UniqueName: \"kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294726 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294757 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.294779 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.403872 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404179 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404211 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404245 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54qjw\" (UniqueName: \"kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404266 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404309 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.404340 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.405397 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.411039 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.419911 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.419922 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.420183 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.420391 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.422600 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54qjw\" (UniqueName: \"kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw\") pod \"barbican-api-5bb7c669fd-grgkd\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.443172 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.660209 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerStarted","Data":"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5"} Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.660508 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.660529 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.693030 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-597c8fcb86-94j25" podStartSLOduration=2.693003244 podStartE2EDuration="2.693003244s" podCreationTimestamp="2025-10-14 08:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:45:15.683995288 +0000 UTC m=+6932.268041915" watchObservedRunningTime="2025-10-14 08:45:15.693003244 +0000 UTC m=+6932.277049871" Oct 14 08:45:15 crc kubenswrapper[5018]: I1014 08:45:15.909116 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 08:45:15 crc kubenswrapper[5018]: W1014 08:45:15.914338 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3c0a334_9cd6_479b_a2c4_c55063d993a2.slice/crio-7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f WatchSource:0}: Error finding container 7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f: Status 404 returned error can't find the container with id 7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.670862 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" event={"ID":"6275743b-3b10-4797-a5f3-ad613eb392f0","Type":"ContainerStarted","Data":"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.671036 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.673514 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerStarted","Data":"652fef03d0063382f0ac1925aeb8888b7617bf80e601c16452dfdee4b721b2c2"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.673539 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerStarted","Data":"c6b8156153448df4ea7ab21d1a8bb3213deefc4aed5b741b5a54c1beb1983b69"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.673550 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerStarted","Data":"7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.673644 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.676262 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerStarted","Data":"f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.676288 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerStarted","Data":"0e6d9ef3c618844329be04c22cf07c96416cade5d7551e947ded7ff89d6b027d"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.678319 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerStarted","Data":"74d5736c20baee1bfbfb8ea1b988e45aea425c479ba0be5662af48c4044d6a13"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.678341 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerStarted","Data":"e8b14272735d2836638396ff83b7ead777ae1c594b519c1e3e4bf3dc34b6cf8b"} Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.695854 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" podStartSLOduration=4.695830697 podStartE2EDuration="4.695830697s" podCreationTimestamp="2025-10-14 08:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:45:16.692045829 +0000 UTC m=+6933.276092466" watchObservedRunningTime="2025-10-14 08:45:16.695830697 +0000 UTC m=+6933.279877334" Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.720155 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" podStartSLOduration=3.228179699 podStartE2EDuration="4.720137929s" podCreationTimestamp="2025-10-14 08:45:12 +0000 UTC" firstStartedPulling="2025-10-14 08:45:13.855229711 +0000 UTC m=+6930.439276328" lastFinishedPulling="2025-10-14 08:45:15.347187931 +0000 UTC m=+6931.931234558" observedRunningTime="2025-10-14 08:45:16.716383872 +0000 UTC m=+6933.300430499" watchObservedRunningTime="2025-10-14 08:45:16.720137929 +0000 UTC m=+6933.304184556" Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.767747 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-854795b57-c4dnm" podStartSLOduration=3.403170825 podStartE2EDuration="4.767729065s" podCreationTimestamp="2025-10-14 08:45:12 +0000 UTC" firstStartedPulling="2025-10-14 08:45:14.002728444 +0000 UTC m=+6930.586775071" lastFinishedPulling="2025-10-14 08:45:15.367286684 +0000 UTC m=+6931.951333311" observedRunningTime="2025-10-14 08:45:16.746925263 +0000 UTC m=+6933.330971890" watchObservedRunningTime="2025-10-14 08:45:16.767729065 +0000 UTC m=+6933.351775692" Oct 14 08:45:16 crc kubenswrapper[5018]: I1014 08:45:16.768817 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5bb7c669fd-grgkd" podStartSLOduration=1.7688109459999999 podStartE2EDuration="1.768810946s" podCreationTimestamp="2025-10-14 08:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:45:16.764244366 +0000 UTC m=+6933.348290993" watchObservedRunningTime="2025-10-14 08:45:16.768810946 +0000 UTC m=+6933.352857573" Oct 14 08:45:17 crc kubenswrapper[5018]: I1014 08:45:17.691443 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.212080 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.302481 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.302872 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="dnsmasq-dns" containerID="cri-o://331ce05842d5501a7cf8976dfe046373e456e44a33387cc341b5bd2004ebe517" gracePeriod=10 Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.758191 5018 generic.go:334] "Generic (PLEG): container finished" podID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerID="331ce05842d5501a7cf8976dfe046373e456e44a33387cc341b5bd2004ebe517" exitCode=0 Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.758882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" event={"ID":"742282bd-dd4d-4cdd-8233-1bd9b9744a36","Type":"ContainerDied","Data":"331ce05842d5501a7cf8976dfe046373e456e44a33387cc341b5bd2004ebe517"} Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.901206 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.979718 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config\") pod \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.979772 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc\") pod \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.979848 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb\") pod \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.979881 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb\") pod \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.979950 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6chw\" (UniqueName: \"kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw\") pod \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\" (UID: \"742282bd-dd4d-4cdd-8233-1bd9b9744a36\") " Oct 14 08:45:23 crc kubenswrapper[5018]: I1014 08:45:23.985788 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw" (OuterVolumeSpecName: "kube-api-access-j6chw") pod "742282bd-dd4d-4cdd-8233-1bd9b9744a36" (UID: "742282bd-dd4d-4cdd-8233-1bd9b9744a36"). InnerVolumeSpecName "kube-api-access-j6chw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.023644 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "742282bd-dd4d-4cdd-8233-1bd9b9744a36" (UID: "742282bd-dd4d-4cdd-8233-1bd9b9744a36"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.033384 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "742282bd-dd4d-4cdd-8233-1bd9b9744a36" (UID: "742282bd-dd4d-4cdd-8233-1bd9b9744a36"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.037192 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config" (OuterVolumeSpecName: "config") pod "742282bd-dd4d-4cdd-8233-1bd9b9744a36" (UID: "742282bd-dd4d-4cdd-8233-1bd9b9744a36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.041096 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "742282bd-dd4d-4cdd-8233-1bd9b9744a36" (UID: "742282bd-dd4d-4cdd-8233-1bd9b9744a36"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.081763 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.081788 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.081798 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.081808 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6chw\" (UniqueName: \"kubernetes.io/projected/742282bd-dd4d-4cdd-8233-1bd9b9744a36-kube-api-access-j6chw\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.081816 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742282bd-dd4d-4cdd-8233-1bd9b9744a36-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.775111 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" event={"ID":"742282bd-dd4d-4cdd-8233-1bd9b9744a36","Type":"ContainerDied","Data":"ae875ee9f4569fd76d831863a64ed9aaa25d1953b20eded5e06275a499f65af0"} Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.775189 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ddcd59f-h5s9f" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.775556 5018 scope.go:117] "RemoveContainer" containerID="331ce05842d5501a7cf8976dfe046373e456e44a33387cc341b5bd2004ebe517" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.799759 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.809170 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746ddcd59f-h5s9f"] Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.836578 5018 scope.go:117] "RemoveContainer" containerID="761be0a3631b10d732cdf168f2d855c2ef99e7a7ef0dedf25c189237a1783d35" Oct 14 08:45:24 crc kubenswrapper[5018]: I1014 08:45:24.983551 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:25 crc kubenswrapper[5018]: I1014 08:45:25.140054 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.616216 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" path="/var/lib/kubelet/pods/742282bd-dd4d-4cdd-8233-1bd9b9744a36/volumes" Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.841896 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.887002 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.979836 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.980105 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-597c8fcb86-94j25" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api-log" containerID="cri-o://b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708" gracePeriod=30 Oct 14 08:45:26 crc kubenswrapper[5018]: I1014 08:45:26.980151 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-597c8fcb86-94j25" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api" containerID="cri-o://3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5" gracePeriod=30 Oct 14 08:45:27 crc kubenswrapper[5018]: I1014 08:45:27.811940 5018 generic.go:334] "Generic (PLEG): container finished" podID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerID="b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708" exitCode=143 Oct 14 08:45:27 crc kubenswrapper[5018]: I1014 08:45:27.812051 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerDied","Data":"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708"} Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.150942 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-597c8fcb86-94j25" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.49:9311/healthcheck\": read tcp 10.217.0.2:33506->10.217.1.49:9311: read: connection reset by peer" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.151076 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-597c8fcb86-94j25" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.49:9311/healthcheck\": read tcp 10.217.0.2:33510->10.217.1.49:9311: read: connection reset by peer" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.599971 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.728290 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data\") pod \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.728369 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnm9n\" (UniqueName: \"kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n\") pod \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.728435 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs\") pod \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.728486 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom\") pod \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.728592 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle\") pod \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\" (UID: \"6a4c3de8-3125-47e2-8c69-bc71222ba3c3\") " Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.729189 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs" (OuterVolumeSpecName: "logs") pod "6a4c3de8-3125-47e2-8c69-bc71222ba3c3" (UID: "6a4c3de8-3125-47e2-8c69-bc71222ba3c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.729779 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.734252 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6a4c3de8-3125-47e2-8c69-bc71222ba3c3" (UID: "6a4c3de8-3125-47e2-8c69-bc71222ba3c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.739572 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n" (OuterVolumeSpecName: "kube-api-access-wnm9n") pod "6a4c3de8-3125-47e2-8c69-bc71222ba3c3" (UID: "6a4c3de8-3125-47e2-8c69-bc71222ba3c3"). InnerVolumeSpecName "kube-api-access-wnm9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.757969 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a4c3de8-3125-47e2-8c69-bc71222ba3c3" (UID: "6a4c3de8-3125-47e2-8c69-bc71222ba3c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.785845 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data" (OuterVolumeSpecName: "config-data") pod "6a4c3de8-3125-47e2-8c69-bc71222ba3c3" (UID: "6a4c3de8-3125-47e2-8c69-bc71222ba3c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.831294 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.831325 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.831336 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.831345 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnm9n\" (UniqueName: \"kubernetes.io/projected/6a4c3de8-3125-47e2-8c69-bc71222ba3c3-kube-api-access-wnm9n\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.850216 5018 generic.go:334] "Generic (PLEG): container finished" podID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerID="3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5" exitCode=0 Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.850249 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerDied","Data":"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5"} Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.850272 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-597c8fcb86-94j25" event={"ID":"6a4c3de8-3125-47e2-8c69-bc71222ba3c3","Type":"ContainerDied","Data":"5061ec448a7bb301146c1e3f2c07d08b38c6be9d52e5c8530876653a768e881a"} Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.850288 5018 scope.go:117] "RemoveContainer" containerID="3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.850300 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-597c8fcb86-94j25" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.896204 5018 scope.go:117] "RemoveContainer" containerID="b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.898263 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.903387 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-597c8fcb86-94j25"] Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.938447 5018 scope.go:117] "RemoveContainer" containerID="3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5" Oct 14 08:45:30 crc kubenswrapper[5018]: E1014 08:45:30.939263 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5\": container with ID starting with 3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5 not found: ID does not exist" containerID="3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.939313 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5"} err="failed to get container status \"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5\": rpc error: code = NotFound desc = could not find container \"3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5\": container with ID starting with 3ecd0df22dcff2fb2e0a9b6d1dcd3718f9d119e9cbf51bbd783c75f409d9d9f5 not found: ID does not exist" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.939343 5018 scope.go:117] "RemoveContainer" containerID="b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708" Oct 14 08:45:30 crc kubenswrapper[5018]: E1014 08:45:30.944281 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708\": container with ID starting with b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708 not found: ID does not exist" containerID="b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708" Oct 14 08:45:30 crc kubenswrapper[5018]: I1014 08:45:30.944370 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708"} err="failed to get container status \"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708\": rpc error: code = NotFound desc = could not find container \"b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708\": container with ID starting with b115113bef81b526518c682afa7ca607db3fc4b02ba2d234f918e1bbaba21708 not found: ID does not exist" Oct 14 08:45:32 crc kubenswrapper[5018]: I1014 08:45:32.624818 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" path="/var/lib/kubelet/pods/6a4c3de8-3125-47e2-8c69-bc71222ba3c3/volumes" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.213608 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6p8hr"] Oct 14 08:45:49 crc kubenswrapper[5018]: E1014 08:45:49.214449 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="init" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214463 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="init" Oct 14 08:45:49 crc kubenswrapper[5018]: E1014 08:45:49.214480 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214487 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api" Oct 14 08:45:49 crc kubenswrapper[5018]: E1014 08:45:49.214502 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="dnsmasq-dns" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214512 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="dnsmasq-dns" Oct 14 08:45:49 crc kubenswrapper[5018]: E1014 08:45:49.214526 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api-log" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214533 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api-log" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214820 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214840 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="742282bd-dd4d-4cdd-8233-1bd9b9744a36" containerName="dnsmasq-dns" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.214851 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4c3de8-3125-47e2-8c69-bc71222ba3c3" containerName="barbican-api-log" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.215501 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.239560 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6p8hr"] Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.341559 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz27m\" (UniqueName: \"kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m\") pod \"neutron-db-create-6p8hr\" (UID: \"77a812d6-06cc-494e-8862-d3d5a4cdef85\") " pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.443406 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz27m\" (UniqueName: \"kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m\") pod \"neutron-db-create-6p8hr\" (UID: \"77a812d6-06cc-494e-8862-d3d5a4cdef85\") " pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.480382 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz27m\" (UniqueName: \"kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m\") pod \"neutron-db-create-6p8hr\" (UID: \"77a812d6-06cc-494e-8862-d3d5a4cdef85\") " pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.544815 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:49 crc kubenswrapper[5018]: I1014 08:45:49.875449 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6p8hr"] Oct 14 08:45:50 crc kubenswrapper[5018]: I1014 08:45:50.049369 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6p8hr" event={"ID":"77a812d6-06cc-494e-8862-d3d5a4cdef85","Type":"ContainerStarted","Data":"47f7114dfd22fca5d153892b0fddff620b178f6206ed69f9ff4e8c54b7a1b3ed"} Oct 14 08:45:50 crc kubenswrapper[5018]: I1014 08:45:50.049414 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6p8hr" event={"ID":"77a812d6-06cc-494e-8862-d3d5a4cdef85","Type":"ContainerStarted","Data":"f2cd0f9bd33c2989bce8d9fb21cb294e6338dc629aed2dc3ce44ba6ddee4735d"} Oct 14 08:45:50 crc kubenswrapper[5018]: I1014 08:45:50.067271 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-6p8hr" podStartSLOduration=1.067252174 podStartE2EDuration="1.067252174s" podCreationTimestamp="2025-10-14 08:45:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:45:50.065019971 +0000 UTC m=+6966.649066598" watchObservedRunningTime="2025-10-14 08:45:50.067252174 +0000 UTC m=+6966.651298811" Oct 14 08:45:51 crc kubenswrapper[5018]: I1014 08:45:51.062686 5018 generic.go:334] "Generic (PLEG): container finished" podID="77a812d6-06cc-494e-8862-d3d5a4cdef85" containerID="47f7114dfd22fca5d153892b0fddff620b178f6206ed69f9ff4e8c54b7a1b3ed" exitCode=0 Oct 14 08:45:51 crc kubenswrapper[5018]: I1014 08:45:51.062852 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6p8hr" event={"ID":"77a812d6-06cc-494e-8862-d3d5a4cdef85","Type":"ContainerDied","Data":"47f7114dfd22fca5d153892b0fddff620b178f6206ed69f9ff4e8c54b7a1b3ed"} Oct 14 08:45:52 crc kubenswrapper[5018]: I1014 08:45:52.411374 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:52 crc kubenswrapper[5018]: I1014 08:45:52.517604 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz27m\" (UniqueName: \"kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m\") pod \"77a812d6-06cc-494e-8862-d3d5a4cdef85\" (UID: \"77a812d6-06cc-494e-8862-d3d5a4cdef85\") " Oct 14 08:45:52 crc kubenswrapper[5018]: I1014 08:45:52.537813 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m" (OuterVolumeSpecName: "kube-api-access-jz27m") pod "77a812d6-06cc-494e-8862-d3d5a4cdef85" (UID: "77a812d6-06cc-494e-8862-d3d5a4cdef85"). InnerVolumeSpecName "kube-api-access-jz27m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:52 crc kubenswrapper[5018]: I1014 08:45:52.619483 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz27m\" (UniqueName: \"kubernetes.io/projected/77a812d6-06cc-494e-8862-d3d5a4cdef85-kube-api-access-jz27m\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:53 crc kubenswrapper[5018]: I1014 08:45:53.085503 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6p8hr" event={"ID":"77a812d6-06cc-494e-8862-d3d5a4cdef85","Type":"ContainerDied","Data":"f2cd0f9bd33c2989bce8d9fb21cb294e6338dc629aed2dc3ce44ba6ddee4735d"} Oct 14 08:45:53 crc kubenswrapper[5018]: I1014 08:45:53.085558 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2cd0f9bd33c2989bce8d9fb21cb294e6338dc629aed2dc3ce44ba6ddee4735d" Oct 14 08:45:53 crc kubenswrapper[5018]: I1014 08:45:53.085610 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6p8hr" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.372383 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d3e5-account-create-pjcwj"] Oct 14 08:45:59 crc kubenswrapper[5018]: E1014 08:45:59.373286 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a812d6-06cc-494e-8862-d3d5a4cdef85" containerName="mariadb-database-create" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.373302 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a812d6-06cc-494e-8862-d3d5a4cdef85" containerName="mariadb-database-create" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.373527 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a812d6-06cc-494e-8862-d3d5a4cdef85" containerName="mariadb-database-create" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.374273 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.376814 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.378390 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d3e5-account-create-pjcwj"] Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.463648 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lpg\" (UniqueName: \"kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg\") pod \"neutron-d3e5-account-create-pjcwj\" (UID: \"de9e6779-e634-4ca2-8724-4b3962dac90b\") " pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.565837 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lpg\" (UniqueName: \"kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg\") pod \"neutron-d3e5-account-create-pjcwj\" (UID: \"de9e6779-e634-4ca2-8724-4b3962dac90b\") " pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.600299 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lpg\" (UniqueName: \"kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg\") pod \"neutron-d3e5-account-create-pjcwj\" (UID: \"de9e6779-e634-4ca2-8724-4b3962dac90b\") " pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:45:59 crc kubenswrapper[5018]: I1014 08:45:59.701967 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:46:00 crc kubenswrapper[5018]: I1014 08:46:00.243919 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d3e5-account-create-pjcwj"] Oct 14 08:46:00 crc kubenswrapper[5018]: W1014 08:46:00.257711 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde9e6779_e634_4ca2_8724_4b3962dac90b.slice/crio-fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44 WatchSource:0}: Error finding container fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44: Status 404 returned error can't find the container with id fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44 Oct 14 08:46:01 crc kubenswrapper[5018]: I1014 08:46:01.161512 5018 generic.go:334] "Generic (PLEG): container finished" podID="de9e6779-e634-4ca2-8724-4b3962dac90b" containerID="ed39e92cd28c36a04d30e0fa9f9aa9bd7b1e722ce24684bfbdf53b76ada30c21" exitCode=0 Oct 14 08:46:01 crc kubenswrapper[5018]: I1014 08:46:01.161579 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3e5-account-create-pjcwj" event={"ID":"de9e6779-e634-4ca2-8724-4b3962dac90b","Type":"ContainerDied","Data":"ed39e92cd28c36a04d30e0fa9f9aa9bd7b1e722ce24684bfbdf53b76ada30c21"} Oct 14 08:46:01 crc kubenswrapper[5018]: I1014 08:46:01.161876 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3e5-account-create-pjcwj" event={"ID":"de9e6779-e634-4ca2-8724-4b3962dac90b","Type":"ContainerStarted","Data":"fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44"} Oct 14 08:46:02 crc kubenswrapper[5018]: I1014 08:46:02.551865 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:46:02 crc kubenswrapper[5018]: I1014 08:46:02.624557 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9lpg\" (UniqueName: \"kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg\") pod \"de9e6779-e634-4ca2-8724-4b3962dac90b\" (UID: \"de9e6779-e634-4ca2-8724-4b3962dac90b\") " Oct 14 08:46:02 crc kubenswrapper[5018]: I1014 08:46:02.635247 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg" (OuterVolumeSpecName: "kube-api-access-t9lpg") pod "de9e6779-e634-4ca2-8724-4b3962dac90b" (UID: "de9e6779-e634-4ca2-8724-4b3962dac90b"). InnerVolumeSpecName "kube-api-access-t9lpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:46:02 crc kubenswrapper[5018]: I1014 08:46:02.730528 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9lpg\" (UniqueName: \"kubernetes.io/projected/de9e6779-e634-4ca2-8724-4b3962dac90b-kube-api-access-t9lpg\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:03 crc kubenswrapper[5018]: I1014 08:46:03.182768 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3e5-account-create-pjcwj" event={"ID":"de9e6779-e634-4ca2-8724-4b3962dac90b","Type":"ContainerDied","Data":"fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44"} Oct 14 08:46:03 crc kubenswrapper[5018]: I1014 08:46:03.182832 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbac12a04ec8df5ade8800f3a0dde4a276a8aec1b5837f9b3074ce2039374e44" Oct 14 08:46:03 crc kubenswrapper[5018]: I1014 08:46:03.183215 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3e5-account-create-pjcwj" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.633939 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-6l49x"] Oct 14 08:46:04 crc kubenswrapper[5018]: E1014 08:46:04.636805 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9e6779-e634-4ca2-8724-4b3962dac90b" containerName="mariadb-account-create" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.638283 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9e6779-e634-4ca2-8724-4b3962dac90b" containerName="mariadb-account-create" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.638883 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9e6779-e634-4ca2-8724-4b3962dac90b" containerName="mariadb-account-create" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.639712 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.641767 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r99t7" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.643082 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.645046 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.647052 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6l49x"] Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.768224 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.768305 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbfhq\" (UniqueName: \"kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.768498 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.869677 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.870079 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.870126 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbfhq\" (UniqueName: \"kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.877693 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.879275 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.888263 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbfhq\" (UniqueName: \"kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq\") pod \"neutron-db-sync-6l49x\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:04 crc kubenswrapper[5018]: I1014 08:46:04.965760 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:05 crc kubenswrapper[5018]: I1014 08:46:05.458959 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6l49x"] Oct 14 08:46:06 crc kubenswrapper[5018]: I1014 08:46:06.225246 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6l49x" event={"ID":"9024d8b2-89ef-4304-9eb7-5d684864402c","Type":"ContainerStarted","Data":"57f99d2fd42f8b9bef5c70891072754cba8a507005f9e58d49af669989298c89"} Oct 14 08:46:06 crc kubenswrapper[5018]: I1014 08:46:06.225565 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6l49x" event={"ID":"9024d8b2-89ef-4304-9eb7-5d684864402c","Type":"ContainerStarted","Data":"54a243bd9836f19489b8e21ae0446acb674f83add2f6af1ff98337c847f72dc8"} Oct 14 08:46:06 crc kubenswrapper[5018]: I1014 08:46:06.260004 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-6l49x" podStartSLOduration=2.2599766470000002 podStartE2EDuration="2.259976647s" podCreationTimestamp="2025-10-14 08:46:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:46:06.254047379 +0000 UTC m=+6982.838094006" watchObservedRunningTime="2025-10-14 08:46:06.259976647 +0000 UTC m=+6982.844023314" Oct 14 08:46:10 crc kubenswrapper[5018]: I1014 08:46:10.267239 5018 generic.go:334] "Generic (PLEG): container finished" podID="9024d8b2-89ef-4304-9eb7-5d684864402c" containerID="57f99d2fd42f8b9bef5c70891072754cba8a507005f9e58d49af669989298c89" exitCode=0 Oct 14 08:46:10 crc kubenswrapper[5018]: I1014 08:46:10.267370 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6l49x" event={"ID":"9024d8b2-89ef-4304-9eb7-5d684864402c","Type":"ContainerDied","Data":"57f99d2fd42f8b9bef5c70891072754cba8a507005f9e58d49af669989298c89"} Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.687895 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.813232 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config\") pod \"9024d8b2-89ef-4304-9eb7-5d684864402c\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.813366 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbfhq\" (UniqueName: \"kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq\") pod \"9024d8b2-89ef-4304-9eb7-5d684864402c\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.813537 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle\") pod \"9024d8b2-89ef-4304-9eb7-5d684864402c\" (UID: \"9024d8b2-89ef-4304-9eb7-5d684864402c\") " Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.820749 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq" (OuterVolumeSpecName: "kube-api-access-xbfhq") pod "9024d8b2-89ef-4304-9eb7-5d684864402c" (UID: "9024d8b2-89ef-4304-9eb7-5d684864402c"). InnerVolumeSpecName "kube-api-access-xbfhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.866158 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config" (OuterVolumeSpecName: "config") pod "9024d8b2-89ef-4304-9eb7-5d684864402c" (UID: "9024d8b2-89ef-4304-9eb7-5d684864402c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.870694 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9024d8b2-89ef-4304-9eb7-5d684864402c" (UID: "9024d8b2-89ef-4304-9eb7-5d684864402c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.918838 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.918881 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbfhq\" (UniqueName: \"kubernetes.io/projected/9024d8b2-89ef-4304-9eb7-5d684864402c-kube-api-access-xbfhq\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:11 crc kubenswrapper[5018]: I1014 08:46:11.918926 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024d8b2-89ef-4304-9eb7-5d684864402c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.296824 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6l49x" event={"ID":"9024d8b2-89ef-4304-9eb7-5d684864402c","Type":"ContainerDied","Data":"54a243bd9836f19489b8e21ae0446acb674f83add2f6af1ff98337c847f72dc8"} Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.296879 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54a243bd9836f19489b8e21ae0446acb674f83add2f6af1ff98337c847f72dc8" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.296881 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6l49x" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.444203 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:46:12 crc kubenswrapper[5018]: E1014 08:46:12.444884 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9024d8b2-89ef-4304-9eb7-5d684864402c" containerName="neutron-db-sync" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.444902 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9024d8b2-89ef-4304-9eb7-5d684864402c" containerName="neutron-db-sync" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.445112 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9024d8b2-89ef-4304-9eb7-5d684864402c" containerName="neutron-db-sync" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.446367 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.462904 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.534230 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.534278 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.534331 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc6h5\" (UniqueName: \"kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.534405 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.534446 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.569628 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.570863 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.575782 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.576043 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.575782 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.576219 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r99t7" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.585758 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.636240 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637099 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637200 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637278 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637354 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637427 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637502 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbrf\" (UniqueName: \"kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637570 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc6h5\" (UniqueName: \"kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637663 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.637740 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.638702 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.639295 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.639930 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.640654 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.660530 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc6h5\" (UniqueName: \"kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5\") pod \"dnsmasq-dns-65449dbc49-tvkxz\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.739761 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbrf\" (UniqueName: \"kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.739884 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.739959 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.739986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.740014 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.745873 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.746451 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.746481 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.745979 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.762960 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.771324 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbrf\" (UniqueName: \"kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf\") pod \"neutron-6559b4ff9b-mwwtw\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:12 crc kubenswrapper[5018]: I1014 08:46:12.892997 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:13 crc kubenswrapper[5018]: I1014 08:46:13.248221 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:46:13 crc kubenswrapper[5018]: I1014 08:46:13.323251 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" event={"ID":"31686a33-209a-4943-bcff-c4a412122856","Type":"ContainerStarted","Data":"14d21006c8281864d3d48149443f800e2af99e41de173325cf73b1da71524b02"} Oct 14 08:46:13 crc kubenswrapper[5018]: I1014 08:46:13.489714 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:13 crc kubenswrapper[5018]: W1014 08:46:13.509288 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod355410bf_f830_48fe_98f2_7cc510d8168b.slice/crio-9ea9edbe7aeeea02fc28e97f1a4cdebb68429cd7e717350ac691e4591a1a2beb WatchSource:0}: Error finding container 9ea9edbe7aeeea02fc28e97f1a4cdebb68429cd7e717350ac691e4591a1a2beb: Status 404 returned error can't find the container with id 9ea9edbe7aeeea02fc28e97f1a4cdebb68429cd7e717350ac691e4591a1a2beb Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.332533 5018 generic.go:334] "Generic (PLEG): container finished" podID="31686a33-209a-4943-bcff-c4a412122856" containerID="eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132" exitCode=0 Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.332665 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" event={"ID":"31686a33-209a-4943-bcff-c4a412122856","Type":"ContainerDied","Data":"eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132"} Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.335038 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerStarted","Data":"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82"} Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.335339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerStarted","Data":"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407"} Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.335466 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerStarted","Data":"9ea9edbe7aeeea02fc28e97f1a4cdebb68429cd7e717350ac691e4591a1a2beb"} Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.335593 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:14 crc kubenswrapper[5018]: I1014 08:46:14.379377 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6559b4ff9b-mwwtw" podStartSLOduration=2.379359339 podStartE2EDuration="2.379359339s" podCreationTimestamp="2025-10-14 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:46:14.378259597 +0000 UTC m=+6990.962306244" watchObservedRunningTime="2025-10-14 08:46:14.379359339 +0000 UTC m=+6990.963405986" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.239167 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.241329 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.244310 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.245512 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.254143 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.286507 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.286973 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v4ms\" (UniqueName: \"kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.287201 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.287419 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.287560 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.287772 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.287982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.347657 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" event={"ID":"31686a33-209a-4943-bcff-c4a412122856","Type":"ContainerStarted","Data":"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f"} Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.348427 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.379915 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" podStartSLOduration=3.379889401 podStartE2EDuration="3.379889401s" podCreationTimestamp="2025-10-14 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:46:15.373087666 +0000 UTC m=+6991.957134303" watchObservedRunningTime="2025-10-14 08:46:15.379889401 +0000 UTC m=+6991.963936028" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389596 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389658 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389684 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389735 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389788 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389818 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v4ms\" (UniqueName: \"kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.389851 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.395580 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.396517 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.397229 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.398061 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.398586 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.409283 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v4ms\" (UniqueName: \"kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.409447 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config\") pod \"neutron-69dd4dcdf-tjqrh\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:15 crc kubenswrapper[5018]: I1014 08:46:15.587011 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:16 crc kubenswrapper[5018]: I1014 08:46:16.141478 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 08:46:16 crc kubenswrapper[5018]: W1014 08:46:16.146371 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1598984_ef08_45c2_a286_fd527ca31a7d.slice/crio-fa6fa0270e513ee2cd3cd6af8513862162036f0f4c655f09b390f49042708c52 WatchSource:0}: Error finding container fa6fa0270e513ee2cd3cd6af8513862162036f0f4c655f09b390f49042708c52: Status 404 returned error can't find the container with id fa6fa0270e513ee2cd3cd6af8513862162036f0f4c655f09b390f49042708c52 Oct 14 08:46:16 crc kubenswrapper[5018]: I1014 08:46:16.368328 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerStarted","Data":"fa6fa0270e513ee2cd3cd6af8513862162036f0f4c655f09b390f49042708c52"} Oct 14 08:46:17 crc kubenswrapper[5018]: I1014 08:46:17.379201 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerStarted","Data":"b153baeb0241dacb32e954510d08cddc263a2d3ee9e254846604179d4e462655"} Oct 14 08:46:17 crc kubenswrapper[5018]: I1014 08:46:17.379506 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerStarted","Data":"645fd93c4abe65ac53978d4f7400e542a575a47bf0b11107c5f6b7d78f3348dd"} Oct 14 08:46:17 crc kubenswrapper[5018]: I1014 08:46:17.379550 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:17 crc kubenswrapper[5018]: I1014 08:46:17.408535 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69dd4dcdf-tjqrh" podStartSLOduration=2.408518686 podStartE2EDuration="2.408518686s" podCreationTimestamp="2025-10-14 08:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:46:17.406457516 +0000 UTC m=+6993.990504203" watchObservedRunningTime="2025-10-14 08:46:17.408518686 +0000 UTC m=+6993.992565313" Oct 14 08:46:22 crc kubenswrapper[5018]: I1014 08:46:22.764891 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:46:22 crc kubenswrapper[5018]: I1014 08:46:22.854183 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:46:22 crc kubenswrapper[5018]: I1014 08:46:22.854641 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="dnsmasq-dns" containerID="cri-o://37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246" gracePeriod=10 Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.406497 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.448038 5018 generic.go:334] "Generic (PLEG): container finished" podID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerID="37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246" exitCode=0 Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.448083 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" event={"ID":"6275743b-3b10-4797-a5f3-ad613eb392f0","Type":"ContainerDied","Data":"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246"} Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.448114 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" event={"ID":"6275743b-3b10-4797-a5f3-ad613eb392f0","Type":"ContainerDied","Data":"01177809583a1274a9098bd68bffca2282ddbe7ebad8de5bb17eeb652e175a1d"} Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.448133 5018 scope.go:117] "RemoveContainer" containerID="37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.448264 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.476593 5018 scope.go:117] "RemoveContainer" containerID="3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.497883 5018 scope.go:117] "RemoveContainer" containerID="37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246" Oct 14 08:46:23 crc kubenswrapper[5018]: E1014 08:46:23.499656 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246\": container with ID starting with 37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246 not found: ID does not exist" containerID="37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.499694 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246"} err="failed to get container status \"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246\": rpc error: code = NotFound desc = could not find container \"37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246\": container with ID starting with 37586c8aba3d7c68c6c89f3a68a1aa67cd9246f0887a0bd58248d088e045a246 not found: ID does not exist" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.499723 5018 scope.go:117] "RemoveContainer" containerID="3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112" Oct 14 08:46:23 crc kubenswrapper[5018]: E1014 08:46:23.500105 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112\": container with ID starting with 3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112 not found: ID does not exist" containerID="3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.500137 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112"} err="failed to get container status \"3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112\": rpc error: code = NotFound desc = could not find container \"3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112\": container with ID starting with 3bc7c12ef7e12bc16507fc9fa621b4c61a536d114035da88e542e6587f9c3112 not found: ID does not exist" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.600322 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb\") pod \"6275743b-3b10-4797-a5f3-ad613eb392f0\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.600403 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb\") pod \"6275743b-3b10-4797-a5f3-ad613eb392f0\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.600464 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wck4\" (UniqueName: \"kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4\") pod \"6275743b-3b10-4797-a5f3-ad613eb392f0\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.600514 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config\") pod \"6275743b-3b10-4797-a5f3-ad613eb392f0\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.600559 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc\") pod \"6275743b-3b10-4797-a5f3-ad613eb392f0\" (UID: \"6275743b-3b10-4797-a5f3-ad613eb392f0\") " Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.614966 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4" (OuterVolumeSpecName: "kube-api-access-2wck4") pod "6275743b-3b10-4797-a5f3-ad613eb392f0" (UID: "6275743b-3b10-4797-a5f3-ad613eb392f0"). InnerVolumeSpecName "kube-api-access-2wck4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.640504 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6275743b-3b10-4797-a5f3-ad613eb392f0" (UID: "6275743b-3b10-4797-a5f3-ad613eb392f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.640895 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6275743b-3b10-4797-a5f3-ad613eb392f0" (UID: "6275743b-3b10-4797-a5f3-ad613eb392f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.641367 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6275743b-3b10-4797-a5f3-ad613eb392f0" (UID: "6275743b-3b10-4797-a5f3-ad613eb392f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.660475 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config" (OuterVolumeSpecName: "config") pod "6275743b-3b10-4797-a5f3-ad613eb392f0" (UID: "6275743b-3b10-4797-a5f3-ad613eb392f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.702792 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.702828 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.702842 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wck4\" (UniqueName: \"kubernetes.io/projected/6275743b-3b10-4797-a5f3-ad613eb392f0-kube-api-access-2wck4\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.702858 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.702869 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6275743b-3b10-4797-a5f3-ad613eb392f0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.785582 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:46:23 crc kubenswrapper[5018]: I1014 08:46:23.791006 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5977c8f7d7-b9c46"] Oct 14 08:46:24 crc kubenswrapper[5018]: I1014 08:46:24.635293 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" path="/var/lib/kubelet/pods/6275743b-3b10-4797-a5f3-ad613eb392f0/volumes" Oct 14 08:46:28 crc kubenswrapper[5018]: I1014 08:46:28.210311 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5977c8f7d7-b9c46" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.46:5353: i/o timeout" Oct 14 08:46:32 crc kubenswrapper[5018]: I1014 08:46:32.463561 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:46:32 crc kubenswrapper[5018]: I1014 08:46:32.463990 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:46:42 crc kubenswrapper[5018]: I1014 08:46:42.909571 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:45 crc kubenswrapper[5018]: I1014 08:46:45.609904 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 08:46:45 crc kubenswrapper[5018]: I1014 08:46:45.712246 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:45 crc kubenswrapper[5018]: I1014 08:46:45.712497 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6559b4ff9b-mwwtw" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-api" containerID="cri-o://b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407" gracePeriod=30 Oct 14 08:46:45 crc kubenswrapper[5018]: I1014 08:46:45.712601 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6559b4ff9b-mwwtw" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-httpd" containerID="cri-o://f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82" gracePeriod=30 Oct 14 08:46:46 crc kubenswrapper[5018]: I1014 08:46:46.709067 5018 generic.go:334] "Generic (PLEG): container finished" podID="355410bf-f830-48fe-98f2-7cc510d8168b" containerID="f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82" exitCode=0 Oct 14 08:46:46 crc kubenswrapper[5018]: I1014 08:46:46.709160 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerDied","Data":"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82"} Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.713528 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.742286 5018 generic.go:334] "Generic (PLEG): container finished" podID="355410bf-f830-48fe-98f2-7cc510d8168b" containerID="b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407" exitCode=0 Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.742340 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerDied","Data":"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407"} Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.742370 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6559b4ff9b-mwwtw" event={"ID":"355410bf-f830-48fe-98f2-7cc510d8168b","Type":"ContainerDied","Data":"9ea9edbe7aeeea02fc28e97f1a4cdebb68429cd7e717350ac691e4591a1a2beb"} Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.742393 5018 scope.go:117] "RemoveContainer" containerID="f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.742542 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6559b4ff9b-mwwtw" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.776064 5018 scope.go:117] "RemoveContainer" containerID="b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.794811 5018 scope.go:117] "RemoveContainer" containerID="f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82" Oct 14 08:46:48 crc kubenswrapper[5018]: E1014 08:46:48.795368 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82\": container with ID starting with f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82 not found: ID does not exist" containerID="f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.795424 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82"} err="failed to get container status \"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82\": rpc error: code = NotFound desc = could not find container \"f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82\": container with ID starting with f131609306881d2369ae6354dba4f46127f124e308810ff0cea978d8c42cad82 not found: ID does not exist" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.795453 5018 scope.go:117] "RemoveContainer" containerID="b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407" Oct 14 08:46:48 crc kubenswrapper[5018]: E1014 08:46:48.795946 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407\": container with ID starting with b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407 not found: ID does not exist" containerID="b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.795985 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407"} err="failed to get container status \"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407\": rpc error: code = NotFound desc = could not find container \"b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407\": container with ID starting with b111c29037529170576a13406df0f372c7ec351bc9bd90a96a546c03b28de407 not found: ID does not exist" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.821679 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs\") pod \"355410bf-f830-48fe-98f2-7cc510d8168b\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.821731 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config\") pod \"355410bf-f830-48fe-98f2-7cc510d8168b\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.821797 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle\") pod \"355410bf-f830-48fe-98f2-7cc510d8168b\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.821870 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbrf\" (UniqueName: \"kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf\") pod \"355410bf-f830-48fe-98f2-7cc510d8168b\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.821921 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config\") pod \"355410bf-f830-48fe-98f2-7cc510d8168b\" (UID: \"355410bf-f830-48fe-98f2-7cc510d8168b\") " Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.828062 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "355410bf-f830-48fe-98f2-7cc510d8168b" (UID: "355410bf-f830-48fe-98f2-7cc510d8168b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.828553 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf" (OuterVolumeSpecName: "kube-api-access-mfbrf") pod "355410bf-f830-48fe-98f2-7cc510d8168b" (UID: "355410bf-f830-48fe-98f2-7cc510d8168b"). InnerVolumeSpecName "kube-api-access-mfbrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.878615 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "355410bf-f830-48fe-98f2-7cc510d8168b" (UID: "355410bf-f830-48fe-98f2-7cc510d8168b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.880792 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config" (OuterVolumeSpecName: "config") pod "355410bf-f830-48fe-98f2-7cc510d8168b" (UID: "355410bf-f830-48fe-98f2-7cc510d8168b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.909126 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "355410bf-f830-48fe-98f2-7cc510d8168b" (UID: "355410bf-f830-48fe-98f2-7cc510d8168b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.924496 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.924538 5018 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.924557 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.924570 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355410bf-f830-48fe-98f2-7cc510d8168b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:48 crc kubenswrapper[5018]: I1014 08:46:48.924587 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbrf\" (UniqueName: \"kubernetes.io/projected/355410bf-f830-48fe-98f2-7cc510d8168b-kube-api-access-mfbrf\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:49 crc kubenswrapper[5018]: I1014 08:46:49.097887 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:49 crc kubenswrapper[5018]: I1014 08:46:49.110531 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6559b4ff9b-mwwtw"] Oct 14 08:46:50 crc kubenswrapper[5018]: I1014 08:46:50.627249 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" path="/var/lib/kubelet/pods/355410bf-f830-48fe-98f2-7cc510d8168b/volumes" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.592434 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-52cnz"] Oct 14 08:47:00 crc kubenswrapper[5018]: E1014 08:47:00.593354 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-api" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593368 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-api" Oct 14 08:47:00 crc kubenswrapper[5018]: E1014 08:47:00.593393 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-httpd" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593400 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-httpd" Oct 14 08:47:00 crc kubenswrapper[5018]: E1014 08:47:00.593412 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="init" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593420 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="init" Oct 14 08:47:00 crc kubenswrapper[5018]: E1014 08:47:00.593432 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="dnsmasq-dns" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593437 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="dnsmasq-dns" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593575 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-api" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593583 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="355410bf-f830-48fe-98f2-7cc510d8168b" containerName="neutron-httpd" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.593602 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6275743b-3b10-4797-a5f3-ad613eb392f0" containerName="dnsmasq-dns" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.594189 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.602559 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.603052 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.603237 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-c2szl" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.618800 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.619405 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.636684 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-52cnz"] Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.742637 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.753105 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.787888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.787990 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.788083 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.788187 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.788248 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bld44\" (UniqueName: \"kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.788331 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.788431 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.801165 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.889844 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.889902 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.889927 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.889954 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.889996 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bld44\" (UniqueName: \"kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890017 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890036 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890060 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890078 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b25rb\" (UniqueName: \"kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890129 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890149 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890183 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.890982 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.891041 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.891332 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.897648 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.900774 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.906052 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.909373 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bld44\" (UniqueName: \"kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44\") pod \"swift-ring-rebalance-52cnz\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.917883 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.992378 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.992846 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b25rb\" (UniqueName: \"kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.992916 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.993022 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.993091 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.994000 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.994270 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.994303 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:00 crc kubenswrapper[5018]: I1014 08:47:00.994751 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.013228 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b25rb\" (UniqueName: \"kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb\") pod \"dnsmasq-dns-8569859d79-xs6xq\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.099176 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.404658 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-52cnz"] Oct 14 08:47:01 crc kubenswrapper[5018]: W1014 08:47:01.416388 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cd8ec80_6c1d_4893_8d7c_765d6966d9ec.slice/crio-56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507 WatchSource:0}: Error finding container 56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507: Status 404 returned error can't find the container with id 56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507 Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.420100 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.525562 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:47:01 crc kubenswrapper[5018]: W1014 08:47:01.531047 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f7c960f_3c71_4fe3_b5f2_c5fe7296464b.slice/crio-07674e4671014f35ecf72a6cb1338cfe65072695af02d0390b1793aff3ec495c WatchSource:0}: Error finding container 07674e4671014f35ecf72a6cb1338cfe65072695af02d0390b1793aff3ec495c: Status 404 returned error can't find the container with id 07674e4671014f35ecf72a6cb1338cfe65072695af02d0390b1793aff3ec495c Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.892538 5018 generic.go:334] "Generic (PLEG): container finished" podID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerID="9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2" exitCode=0 Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.892607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" event={"ID":"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b","Type":"ContainerDied","Data":"9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2"} Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.892907 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" event={"ID":"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b","Type":"ContainerStarted","Data":"07674e4671014f35ecf72a6cb1338cfe65072695af02d0390b1793aff3ec495c"} Oct 14 08:47:01 crc kubenswrapper[5018]: I1014 08:47:01.894456 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-52cnz" event={"ID":"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec","Type":"ContainerStarted","Data":"56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507"} Oct 14 08:47:02 crc kubenswrapper[5018]: I1014 08:47:02.463295 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:47:02 crc kubenswrapper[5018]: I1014 08:47:02.463350 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:47:02 crc kubenswrapper[5018]: I1014 08:47:02.905172 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" event={"ID":"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b","Type":"ContainerStarted","Data":"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb"} Oct 14 08:47:02 crc kubenswrapper[5018]: I1014 08:47:02.905664 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:02 crc kubenswrapper[5018]: I1014 08:47:02.927525 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" podStartSLOduration=2.927509727 podStartE2EDuration="2.927509727s" podCreationTimestamp="2025-10-14 08:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:47:02.924436509 +0000 UTC m=+7039.508483136" watchObservedRunningTime="2025-10-14 08:47:02.927509727 +0000 UTC m=+7039.511556344" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.124207 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.125478 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.127561 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.140814 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238412 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238468 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238497 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238590 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sx97\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238794 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.238870 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.340973 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.341035 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.341071 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.341126 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sx97\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.341168 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.341195 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.342273 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.342463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.348424 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.348701 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.350493 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.363282 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sx97\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97\") pod \"swift-proxy-656b97cf6d-wm6kk\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:03 crc kubenswrapper[5018]: I1014 08:47:03.448641 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.407044 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.408782 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.423129 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.445122 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.445581 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.558721 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.558846 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.558965 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.558993 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xjbm\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.559037 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.559185 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.559237 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.559426 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661572 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661670 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661712 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661731 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xjbm\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661776 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661811 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661844 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.661871 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.662673 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.663059 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.667287 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.670750 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.679315 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xjbm\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.679481 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.681720 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.683499 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs\") pod \"swift-proxy-559f8b448b-zgvh5\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:04 crc kubenswrapper[5018]: I1014 08:47:04.764465 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.218310 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 08:47:05 crc kubenswrapper[5018]: W1014 08:47:05.219797 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod521da8cb_e7fb_42b0_aa61_ee86ec329fad.slice/crio-cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9 WatchSource:0}: Error finding container cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9: Status 404 returned error can't find the container with id cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9 Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.943724 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerStarted","Data":"f3bfacf0b50c0a9351a97df8cd9a4d4d20587d8f3b2250add56a2512071e49f6"} Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.944746 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerStarted","Data":"8ea98fcf3c27a68c727221403d77e0a9e5e402b4f567534f6bb06689161fc3eb"} Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.944795 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerStarted","Data":"cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9"} Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.944876 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.944914 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.947666 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-52cnz" event={"ID":"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec","Type":"ContainerStarted","Data":"377bff2c1f590ed92d30d5e19176b01620ec07050618bc1c81fbd84b01fd16ca"} Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.984444 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-559f8b448b-zgvh5" podStartSLOduration=1.98442193 podStartE2EDuration="1.98442193s" podCreationTimestamp="2025-10-14 08:47:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:47:05.968997518 +0000 UTC m=+7042.553044145" watchObservedRunningTime="2025-10-14 08:47:05.98442193 +0000 UTC m=+7042.568468557" Oct 14 08:47:05 crc kubenswrapper[5018]: I1014 08:47:05.994406 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-52cnz" podStartSLOduration=2.656041504 podStartE2EDuration="5.994376916s" podCreationTimestamp="2025-10-14 08:47:00 +0000 UTC" firstStartedPulling="2025-10-14 08:47:01.419827906 +0000 UTC m=+7038.003874533" lastFinishedPulling="2025-10-14 08:47:04.758163318 +0000 UTC m=+7041.342209945" observedRunningTime="2025-10-14 08:47:05.986289234 +0000 UTC m=+7042.570335881" watchObservedRunningTime="2025-10-14 08:47:05.994376916 +0000 UTC m=+7042.578423583" Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.221042 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:06 crc kubenswrapper[5018]: W1014 08:47:06.235397 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4f4a5e0_4f76_485c_85d0_0c49f85d98f1.slice/crio-7b8460f7b5f2a879bb19ab201ebcc243efde24dd632ca4361ced210cca6cbeb3 WatchSource:0}: Error finding container 7b8460f7b5f2a879bb19ab201ebcc243efde24dd632ca4361ced210cca6cbeb3: Status 404 returned error can't find the container with id 7b8460f7b5f2a879bb19ab201ebcc243efde24dd632ca4361ced210cca6cbeb3 Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.960828 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerStarted","Data":"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4"} Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.961158 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerStarted","Data":"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04"} Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.961180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerStarted","Data":"7b8460f7b5f2a879bb19ab201ebcc243efde24dd632ca4361ced210cca6cbeb3"} Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.961202 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:06 crc kubenswrapper[5018]: I1014 08:47:06.979982 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-656b97cf6d-wm6kk" podStartSLOduration=3.9799632689999997 podStartE2EDuration="3.979963269s" podCreationTimestamp="2025-10-14 08:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:47:06.976963223 +0000 UTC m=+7043.561009860" watchObservedRunningTime="2025-10-14 08:47:06.979963269 +0000 UTC m=+7043.564009896" Oct 14 08:47:07 crc kubenswrapper[5018]: I1014 08:47:07.979982 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:08 crc kubenswrapper[5018]: I1014 08:47:08.359318 5018 scope.go:117] "RemoveContainer" containerID="d2bd12da785c7782d1bcd148149e961b39f28373632a498c2c01b8d852f3ec54" Oct 14 08:47:08 crc kubenswrapper[5018]: I1014 08:47:08.393892 5018 scope.go:117] "RemoveContainer" containerID="10f130c9b8f9466bd0e5944aaa5001f9528abd0308435ffa589e4683d51b114b" Oct 14 08:47:08 crc kubenswrapper[5018]: I1014 08:47:08.996520 5018 generic.go:334] "Generic (PLEG): container finished" podID="3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" containerID="377bff2c1f590ed92d30d5e19176b01620ec07050618bc1c81fbd84b01fd16ca" exitCode=0 Oct 14 08:47:08 crc kubenswrapper[5018]: I1014 08:47:08.996726 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-52cnz" event={"ID":"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec","Type":"ContainerDied","Data":"377bff2c1f590ed92d30d5e19176b01620ec07050618bc1c81fbd84b01fd16ca"} Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.306482 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.491147 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.492175 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.492344 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.492684 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.493049 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.493408 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.493643 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bld44\" (UniqueName: \"kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44\") pod \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\" (UID: \"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec\") " Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.498271 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.498853 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.502907 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44" (OuterVolumeSpecName: "kube-api-access-bld44") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "kube-api-access-bld44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.520307 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.531145 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.535547 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts" (OuterVolumeSpecName: "scripts") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.541450 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" (UID: "3cd8ec80-6c1d-4893-8d7c-765d6966d9ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597102 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597320 5018 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597419 5018 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597492 5018 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597561 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597648 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bld44\" (UniqueName: \"kubernetes.io/projected/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-kube-api-access-bld44\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:10 crc kubenswrapper[5018]: I1014 08:47:10.597720 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.018452 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-52cnz" event={"ID":"3cd8ec80-6c1d-4893-8d7c-765d6966d9ec","Type":"ContainerDied","Data":"56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507"} Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.018532 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56abf7584a2ac1022a9e92d9846e760bed73cee2fc150627a63d3fa8fc138507" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.018578 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-52cnz" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.102135 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.180792 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.181165 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="dnsmasq-dns" containerID="cri-o://f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f" gracePeriod=10 Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.644352 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.822843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config\") pod \"31686a33-209a-4943-bcff-c4a412122856\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.822880 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc\") pod \"31686a33-209a-4943-bcff-c4a412122856\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.822932 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc6h5\" (UniqueName: \"kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5\") pod \"31686a33-209a-4943-bcff-c4a412122856\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.823036 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb\") pod \"31686a33-209a-4943-bcff-c4a412122856\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.823807 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb\") pod \"31686a33-209a-4943-bcff-c4a412122856\" (UID: \"31686a33-209a-4943-bcff-c4a412122856\") " Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.829856 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5" (OuterVolumeSpecName: "kube-api-access-nc6h5") pod "31686a33-209a-4943-bcff-c4a412122856" (UID: "31686a33-209a-4943-bcff-c4a412122856"). InnerVolumeSpecName "kube-api-access-nc6h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.877951 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "31686a33-209a-4943-bcff-c4a412122856" (UID: "31686a33-209a-4943-bcff-c4a412122856"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.879564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "31686a33-209a-4943-bcff-c4a412122856" (UID: "31686a33-209a-4943-bcff-c4a412122856"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.881534 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config" (OuterVolumeSpecName: "config") pod "31686a33-209a-4943-bcff-c4a412122856" (UID: "31686a33-209a-4943-bcff-c4a412122856"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.897362 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "31686a33-209a-4943-bcff-c4a412122856" (UID: "31686a33-209a-4943-bcff-c4a412122856"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.926099 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.926167 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.926183 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.926226 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc6h5\" (UniqueName: \"kubernetes.io/projected/31686a33-209a-4943-bcff-c4a412122856-kube-api-access-nc6h5\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:11 crc kubenswrapper[5018]: I1014 08:47:11.926243 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31686a33-209a-4943-bcff-c4a412122856-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.054911 5018 generic.go:334] "Generic (PLEG): container finished" podID="31686a33-209a-4943-bcff-c4a412122856" containerID="f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f" exitCode=0 Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.054957 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" event={"ID":"31686a33-209a-4943-bcff-c4a412122856","Type":"ContainerDied","Data":"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f"} Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.054993 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" event={"ID":"31686a33-209a-4943-bcff-c4a412122856","Type":"ContainerDied","Data":"14d21006c8281864d3d48149443f800e2af99e41de173325cf73b1da71524b02"} Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.055011 5018 scope.go:117] "RemoveContainer" containerID="f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.055052 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65449dbc49-tvkxz" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.075435 5018 scope.go:117] "RemoveContainer" containerID="eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.099766 5018 scope.go:117] "RemoveContainer" containerID="f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f" Oct 14 08:47:12 crc kubenswrapper[5018]: E1014 08:47:12.100267 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f\": container with ID starting with f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f not found: ID does not exist" containerID="f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.100311 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f"} err="failed to get container status \"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f\": rpc error: code = NotFound desc = could not find container \"f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f\": container with ID starting with f2712b389fb3928d67524b895c86182ee5fdbf1d0acabd2fcba45f39ceaba01f not found: ID does not exist" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.100339 5018 scope.go:117] "RemoveContainer" containerID="eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132" Oct 14 08:47:12 crc kubenswrapper[5018]: E1014 08:47:12.100679 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132\": container with ID starting with eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132 not found: ID does not exist" containerID="eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.100724 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132"} err="failed to get container status \"eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132\": rpc error: code = NotFound desc = could not find container \"eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132\": container with ID starting with eda2181fcdb0627445f0eae12ada7e0f121b02894a202fecfd6578af8efa2132 not found: ID does not exist" Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.103305 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.109040 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65449dbc49-tvkxz"] Oct 14 08:47:12 crc kubenswrapper[5018]: I1014 08:47:12.624991 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31686a33-209a-4943-bcff-c4a412122856" path="/var/lib/kubelet/pods/31686a33-209a-4943-bcff-c4a412122856/volumes" Oct 14 08:47:13 crc kubenswrapper[5018]: I1014 08:47:13.455357 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:13 crc kubenswrapper[5018]: I1014 08:47:13.457178 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:14 crc kubenswrapper[5018]: I1014 08:47:14.773562 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:14 crc kubenswrapper[5018]: I1014 08:47:14.780579 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 08:47:14 crc kubenswrapper[5018]: I1014 08:47:14.868924 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:14 crc kubenswrapper[5018]: I1014 08:47:14.869491 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-656b97cf6d-wm6kk" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-httpd" containerID="cri-o://3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04" gracePeriod=30 Oct 14 08:47:14 crc kubenswrapper[5018]: I1014 08:47:14.870111 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-656b97cf6d-wm6kk" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-server" containerID="cri-o://bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4" gracePeriod=30 Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.090728 5018 generic.go:334] "Generic (PLEG): container finished" podID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerID="bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4" exitCode=0 Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.091526 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerDied","Data":"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4"} Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.683702 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.829520 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.829874 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.829949 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.829977 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sx97\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.830014 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.830043 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd\") pod \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\" (UID: \"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1\") " Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.830413 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.831021 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.834878 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.840747 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97" (OuterVolumeSpecName: "kube-api-access-9sx97") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "kube-api-access-9sx97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.875892 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.894073 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data" (OuterVolumeSpecName: "config-data") pod "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" (UID: "c4f4a5e0-4f76-485c-85d0-0c49f85d98f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932111 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932146 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932155 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932165 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sx97\" (UniqueName: \"kubernetes.io/projected/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-kube-api-access-9sx97\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932174 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:15 crc kubenswrapper[5018]: I1014 08:47:15.932182 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.105193 5018 generic.go:334] "Generic (PLEG): container finished" podID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerID="3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04" exitCode=0 Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.105235 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerDied","Data":"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04"} Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.105260 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-656b97cf6d-wm6kk" event={"ID":"c4f4a5e0-4f76-485c-85d0-0c49f85d98f1","Type":"ContainerDied","Data":"7b8460f7b5f2a879bb19ab201ebcc243efde24dd632ca4361ced210cca6cbeb3"} Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.105276 5018 scope.go:117] "RemoveContainer" containerID="bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.105276 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-656b97cf6d-wm6kk" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.151716 5018 scope.go:117] "RemoveContainer" containerID="3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.203603 5018 scope.go:117] "RemoveContainer" containerID="bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4" Oct 14 08:47:16 crc kubenswrapper[5018]: E1014 08:47:16.204700 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4\": container with ID starting with bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4 not found: ID does not exist" containerID="bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.204754 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4"} err="failed to get container status \"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4\": rpc error: code = NotFound desc = could not find container \"bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4\": container with ID starting with bcbb84b6737abcf30f78d202857229beea6ae1195c2fa6e1c64af575399c88a4 not found: ID does not exist" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.204783 5018 scope.go:117] "RemoveContainer" containerID="3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04" Oct 14 08:47:16 crc kubenswrapper[5018]: E1014 08:47:16.205430 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04\": container with ID starting with 3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04 not found: ID does not exist" containerID="3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.205472 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04"} err="failed to get container status \"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04\": rpc error: code = NotFound desc = could not find container \"3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04\": container with ID starting with 3879aea5988dd3595ed9341ac9895d629639bf0f89fb7a9509d50b4fe6fe7a04 not found: ID does not exist" Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.211909 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.220089 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-656b97cf6d-wm6kk"] Oct 14 08:47:16 crc kubenswrapper[5018]: I1014 08:47:16.621815 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" path="/var/lib/kubelet/pods/c4f4a5e0-4f76-485c-85d0-0c49f85d98f1/volumes" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.990988 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jchrc"] Oct 14 08:47:20 crc kubenswrapper[5018]: E1014 08:47:20.991695 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="init" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.991712 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="init" Oct 14 08:47:20 crc kubenswrapper[5018]: E1014 08:47:20.991744 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" containerName="swift-ring-rebalance" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.991752 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" containerName="swift-ring-rebalance" Oct 14 08:47:20 crc kubenswrapper[5018]: E1014 08:47:20.991768 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="dnsmasq-dns" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.991775 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="dnsmasq-dns" Oct 14 08:47:20 crc kubenswrapper[5018]: E1014 08:47:20.991805 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-server" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.991813 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-server" Oct 14 08:47:20 crc kubenswrapper[5018]: E1014 08:47:20.991827 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-httpd" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.991835 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-httpd" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.992076 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="31686a33-209a-4943-bcff-c4a412122856" containerName="dnsmasq-dns" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.992093 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" containerName="swift-ring-rebalance" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.992111 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-server" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.992130 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f4a5e0-4f76-485c-85d0-0c49f85d98f1" containerName="proxy-httpd" Oct 14 08:47:20 crc kubenswrapper[5018]: I1014 08:47:20.992793 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.001361 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jchrc"] Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.053393 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk9b9\" (UniqueName: \"kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9\") pod \"cinder-db-create-jchrc\" (UID: \"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22\") " pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.154644 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk9b9\" (UniqueName: \"kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9\") pod \"cinder-db-create-jchrc\" (UID: \"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22\") " pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.185550 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk9b9\" (UniqueName: \"kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9\") pod \"cinder-db-create-jchrc\" (UID: \"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22\") " pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.311601 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:21 crc kubenswrapper[5018]: I1014 08:47:21.822180 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jchrc"] Oct 14 08:47:22 crc kubenswrapper[5018]: I1014 08:47:22.167154 5018 generic.go:334] "Generic (PLEG): container finished" podID="e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" containerID="5ead85004aed5d89f1f3d3206f70176e9cee5c2791b72807af02242104b80b7b" exitCode=0 Oct 14 08:47:22 crc kubenswrapper[5018]: I1014 08:47:22.167194 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jchrc" event={"ID":"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22","Type":"ContainerDied","Data":"5ead85004aed5d89f1f3d3206f70176e9cee5c2791b72807af02242104b80b7b"} Oct 14 08:47:22 crc kubenswrapper[5018]: I1014 08:47:22.167217 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jchrc" event={"ID":"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22","Type":"ContainerStarted","Data":"8369e73c6a829f4e98f589e8873d4f1da4b596febc3d7cf0d5cf445c0b55bc5d"} Oct 14 08:47:23 crc kubenswrapper[5018]: I1014 08:47:23.654830 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:23 crc kubenswrapper[5018]: I1014 08:47:23.805949 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk9b9\" (UniqueName: \"kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9\") pod \"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22\" (UID: \"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22\") " Oct 14 08:47:23 crc kubenswrapper[5018]: I1014 08:47:23.815861 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9" (OuterVolumeSpecName: "kube-api-access-lk9b9") pod "e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" (UID: "e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22"). InnerVolumeSpecName "kube-api-access-lk9b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:23 crc kubenswrapper[5018]: I1014 08:47:23.909344 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk9b9\" (UniqueName: \"kubernetes.io/projected/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22-kube-api-access-lk9b9\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:24 crc kubenswrapper[5018]: I1014 08:47:24.193142 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jchrc" event={"ID":"e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22","Type":"ContainerDied","Data":"8369e73c6a829f4e98f589e8873d4f1da4b596febc3d7cf0d5cf445c0b55bc5d"} Oct 14 08:47:24 crc kubenswrapper[5018]: I1014 08:47:24.193208 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8369e73c6a829f4e98f589e8873d4f1da4b596febc3d7cf0d5cf445c0b55bc5d" Oct 14 08:47:24 crc kubenswrapper[5018]: I1014 08:47:24.193225 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jchrc" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.125019 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3690-account-create-zl9fz"] Oct 14 08:47:31 crc kubenswrapper[5018]: E1014 08:47:31.126300 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" containerName="mariadb-database-create" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.126322 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" containerName="mariadb-database-create" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.126705 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" containerName="mariadb-database-create" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.127798 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.133930 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.142923 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3690-account-create-zl9fz"] Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.179443 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltcnl\" (UniqueName: \"kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl\") pod \"cinder-3690-account-create-zl9fz\" (UID: \"998e8c25-bbdf-43b3-a985-3b4c96e25e10\") " pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.281264 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltcnl\" (UniqueName: \"kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl\") pod \"cinder-3690-account-create-zl9fz\" (UID: \"998e8c25-bbdf-43b3-a985-3b4c96e25e10\") " pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.301531 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltcnl\" (UniqueName: \"kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl\") pod \"cinder-3690-account-create-zl9fz\" (UID: \"998e8c25-bbdf-43b3-a985-3b4c96e25e10\") " pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.467482 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:31 crc kubenswrapper[5018]: I1014 08:47:31.983053 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3690-account-create-zl9fz"] Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.277083 5018 generic.go:334] "Generic (PLEG): container finished" podID="998e8c25-bbdf-43b3-a985-3b4c96e25e10" containerID="17469fb7cbaaba599763c7ea2a81ff458a929f8ed6bb8547cf34cdf69db914eb" exitCode=0 Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.277144 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3690-account-create-zl9fz" event={"ID":"998e8c25-bbdf-43b3-a985-3b4c96e25e10","Type":"ContainerDied","Data":"17469fb7cbaaba599763c7ea2a81ff458a929f8ed6bb8547cf34cdf69db914eb"} Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.277172 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3690-account-create-zl9fz" event={"ID":"998e8c25-bbdf-43b3-a985-3b4c96e25e10","Type":"ContainerStarted","Data":"80ff9bc176bc16b034efb40c417de72ecce39d7f6254f32c26f67c59dfd1798a"} Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.463168 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.463263 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.463331 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.464372 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:47:32 crc kubenswrapper[5018]: I1014 08:47:32.464481 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" gracePeriod=600 Oct 14 08:47:32 crc kubenswrapper[5018]: E1014 08:47:32.597647 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.305656 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" exitCode=0 Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.306184 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44"} Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.306285 5018 scope.go:117] "RemoveContainer" containerID="882d0cdbe59daefdb56680647122f93c4b5f253918af9ab786beb2d0c64114ee" Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.308082 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:47:33 crc kubenswrapper[5018]: E1014 08:47:33.309586 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.703787 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.726612 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltcnl\" (UniqueName: \"kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl\") pod \"998e8c25-bbdf-43b3-a985-3b4c96e25e10\" (UID: \"998e8c25-bbdf-43b3-a985-3b4c96e25e10\") " Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.735504 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl" (OuterVolumeSpecName: "kube-api-access-ltcnl") pod "998e8c25-bbdf-43b3-a985-3b4c96e25e10" (UID: "998e8c25-bbdf-43b3-a985-3b4c96e25e10"). InnerVolumeSpecName "kube-api-access-ltcnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:33 crc kubenswrapper[5018]: I1014 08:47:33.829094 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltcnl\" (UniqueName: \"kubernetes.io/projected/998e8c25-bbdf-43b3-a985-3b4c96e25e10-kube-api-access-ltcnl\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:34 crc kubenswrapper[5018]: I1014 08:47:34.323374 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3690-account-create-zl9fz" event={"ID":"998e8c25-bbdf-43b3-a985-3b4c96e25e10","Type":"ContainerDied","Data":"80ff9bc176bc16b034efb40c417de72ecce39d7f6254f32c26f67c59dfd1798a"} Oct 14 08:47:34 crc kubenswrapper[5018]: I1014 08:47:34.323436 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ff9bc176bc16b034efb40c417de72ecce39d7f6254f32c26f67c59dfd1798a" Oct 14 08:47:34 crc kubenswrapper[5018]: I1014 08:47:34.324918 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3690-account-create-zl9fz" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.296392 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vwvpv"] Oct 14 08:47:36 crc kubenswrapper[5018]: E1014 08:47:36.297059 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998e8c25-bbdf-43b3-a985-3b4c96e25e10" containerName="mariadb-account-create" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.297075 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="998e8c25-bbdf-43b3-a985-3b4c96e25e10" containerName="mariadb-account-create" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.297233 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="998e8c25-bbdf-43b3-a985-3b4c96e25e10" containerName="mariadb-account-create" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.297765 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.300566 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.301377 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.301492 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9lbdk" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.317862 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vwvpv"] Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.380868 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.380913 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz9nm\" (UniqueName: \"kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.380951 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.381027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.381102 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.381119 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483090 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483134 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483184 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483209 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz9nm\" (UniqueName: \"kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483320 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.483327 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.490733 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.491761 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.492768 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.494315 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.500826 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz9nm\" (UniqueName: \"kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm\") pod \"cinder-db-sync-vwvpv\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:36 crc kubenswrapper[5018]: I1014 08:47:36.613440 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:47:37 crc kubenswrapper[5018]: I1014 08:47:37.103088 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vwvpv"] Oct 14 08:47:37 crc kubenswrapper[5018]: I1014 08:47:37.361300 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vwvpv" event={"ID":"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a","Type":"ContainerStarted","Data":"7af295dafeb802d77a585f38dd044d486462579432ef79aa9c3befaeee3a36f3"} Oct 14 08:47:43 crc kubenswrapper[5018]: I1014 08:47:43.606495 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:47:43 crc kubenswrapper[5018]: E1014 08:47:43.607222 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:47:56 crc kubenswrapper[5018]: I1014 08:47:56.606198 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:47:56 crc kubenswrapper[5018]: E1014 08:47:56.607068 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:47:57 crc kubenswrapper[5018]: I1014 08:47:57.585762 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vwvpv" event={"ID":"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a","Type":"ContainerStarted","Data":"555d0de4d32cd68c3cbd905d73c9af6bf2510788a6b940bc0ecdacda4f168243"} Oct 14 08:47:59 crc kubenswrapper[5018]: I1014 08:47:59.611772 5018 generic.go:334] "Generic (PLEG): container finished" podID="f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" containerID="555d0de4d32cd68c3cbd905d73c9af6bf2510788a6b940bc0ecdacda4f168243" exitCode=0 Oct 14 08:47:59 crc kubenswrapper[5018]: I1014 08:47:59.611886 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vwvpv" event={"ID":"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a","Type":"ContainerDied","Data":"555d0de4d32cd68c3cbd905d73c9af6bf2510788a6b940bc0ecdacda4f168243"} Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.060082 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166160 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166399 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166439 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166470 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz9nm\" (UniqueName: \"kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166519 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166581 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data\") pod \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\" (UID: \"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a\") " Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.166563 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.167077 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.175704 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm" (OuterVolumeSpecName: "kube-api-access-dz9nm") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "kube-api-access-dz9nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.175983 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts" (OuterVolumeSpecName: "scripts") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.176830 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.205525 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.228896 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data" (OuterVolumeSpecName: "config-data") pod "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" (UID: "f7d26e3c-22a7-4fa5-8f89-4718f312fe4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.268520 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.268554 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz9nm\" (UniqueName: \"kubernetes.io/projected/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-kube-api-access-dz9nm\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.268931 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.269300 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.269396 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.640585 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vwvpv" event={"ID":"f7d26e3c-22a7-4fa5-8f89-4718f312fe4a","Type":"ContainerDied","Data":"7af295dafeb802d77a585f38dd044d486462579432ef79aa9c3befaeee3a36f3"} Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.641019 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7af295dafeb802d77a585f38dd044d486462579432ef79aa9c3befaeee3a36f3" Oct 14 08:48:01 crc kubenswrapper[5018]: I1014 08:48:01.640705 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vwvpv" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.429469 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:48:02 crc kubenswrapper[5018]: E1014 08:48:02.430256 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" containerName="cinder-db-sync" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.430275 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" containerName="cinder-db-sync" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.430488 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" containerName="cinder-db-sync" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.435828 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.462606 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.498734 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.498827 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8mv6\" (UniqueName: \"kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.498888 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.498922 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.499016 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.599801 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8mv6\" (UniqueName: \"kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.599895 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.599986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.601080 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.601115 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.601325 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.601513 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.602291 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.602299 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.622522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8mv6\" (UniqueName: \"kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6\") pod \"dnsmasq-dns-9c4b98475-w5dsc\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.634049 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.635424 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.637375 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.637566 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.638342 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.638497 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9lbdk" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.651461 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702361 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702477 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702504 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb7kx\" (UniqueName: \"kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702722 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.702905 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.703147 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.754106 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805535 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805718 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805785 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805795 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805811 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb7kx\" (UniqueName: \"kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805896 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.805962 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.806020 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.807313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.812343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.822058 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.822327 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb7kx\" (UniqueName: \"kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.826405 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.826839 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " pod="openstack/cinder-api-0" Oct 14 08:48:02 crc kubenswrapper[5018]: I1014 08:48:02.981606 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.257785 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.504358 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:03 crc kubenswrapper[5018]: W1014 08:48:03.561464 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d1c1369_34a3_48a5_acfb_bb4232a523cb.slice/crio-666d32da5e6c45a1d48d0ef0c76ed76b0c47ad73e5b8a6c0440d75e7ee7b6d64 WatchSource:0}: Error finding container 666d32da5e6c45a1d48d0ef0c76ed76b0c47ad73e5b8a6c0440d75e7ee7b6d64: Status 404 returned error can't find the container with id 666d32da5e6c45a1d48d0ef0c76ed76b0c47ad73e5b8a6c0440d75e7ee7b6d64 Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.657304 5018 generic.go:334] "Generic (PLEG): container finished" podID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerID="cfeed0d822d12409d089b725976adacba4ca89b9c79a4c2f02243af9ead8618b" exitCode=0 Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.657372 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" event={"ID":"56caaa8c-1769-4715-a5e4-ec385d2d6f01","Type":"ContainerDied","Data":"cfeed0d822d12409d089b725976adacba4ca89b9c79a4c2f02243af9ead8618b"} Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.657803 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" event={"ID":"56caaa8c-1769-4715-a5e4-ec385d2d6f01","Type":"ContainerStarted","Data":"9bb629657c7eca9b449456dd9337c4b6e6738dacb5a51e40c4ef51b480fd67ac"} Oct 14 08:48:03 crc kubenswrapper[5018]: I1014 08:48:03.662605 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerStarted","Data":"666d32da5e6c45a1d48d0ef0c76ed76b0c47ad73e5b8a6c0440d75e7ee7b6d64"} Oct 14 08:48:04 crc kubenswrapper[5018]: I1014 08:48:04.543014 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:04 crc kubenswrapper[5018]: I1014 08:48:04.680355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" event={"ID":"56caaa8c-1769-4715-a5e4-ec385d2d6f01","Type":"ContainerStarted","Data":"efd60405ff584a3651531e9f90a1f89f5dcc5641baaff03e8db0c27e7d450878"} Oct 14 08:48:04 crc kubenswrapper[5018]: I1014 08:48:04.681385 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:04 crc kubenswrapper[5018]: I1014 08:48:04.689077 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerStarted","Data":"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1"} Oct 14 08:48:04 crc kubenswrapper[5018]: I1014 08:48:04.707798 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" podStartSLOduration=2.707777319 podStartE2EDuration="2.707777319s" podCreationTimestamp="2025-10-14 08:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:48:04.703344732 +0000 UTC m=+7101.287391359" watchObservedRunningTime="2025-10-14 08:48:04.707777319 +0000 UTC m=+7101.291823946" Oct 14 08:48:05 crc kubenswrapper[5018]: I1014 08:48:05.701388 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerStarted","Data":"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e"} Oct 14 08:48:05 crc kubenswrapper[5018]: I1014 08:48:05.701764 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api-log" containerID="cri-o://9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" gracePeriod=30 Oct 14 08:48:05 crc kubenswrapper[5018]: I1014 08:48:05.702203 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api" containerID="cri-o://5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" gracePeriod=30 Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.326413 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493668 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493782 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493804 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493800 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493832 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493905 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.493962 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb7kx\" (UniqueName: \"kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.494016 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom\") pod \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\" (UID: \"2d1c1369-34a3-48a5-acfb-bb4232a523cb\") " Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.494321 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d1c1369-34a3-48a5-acfb-bb4232a523cb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.494313 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs" (OuterVolumeSpecName: "logs") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.499391 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.502881 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts" (OuterVolumeSpecName: "scripts") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.505904 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx" (OuterVolumeSpecName: "kube-api-access-sb7kx") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "kube-api-access-sb7kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.540366 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.552680 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data" (OuterVolumeSpecName: "config-data") pod "2d1c1369-34a3-48a5-acfb-bb4232a523cb" (UID: "2d1c1369-34a3-48a5-acfb-bb4232a523cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596801 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb7kx\" (UniqueName: \"kubernetes.io/projected/2d1c1369-34a3-48a5-acfb-bb4232a523cb-kube-api-access-sb7kx\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596864 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596887 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d1c1369-34a3-48a5-acfb-bb4232a523cb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596912 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596939 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.596963 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d1c1369-34a3-48a5-acfb-bb4232a523cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713783 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerID="5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" exitCode=0 Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713816 5018 generic.go:334] "Generic (PLEG): container finished" podID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerID="9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" exitCode=143 Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713836 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerDied","Data":"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e"} Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713896 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerDied","Data":"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1"} Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713913 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2d1c1369-34a3-48a5-acfb-bb4232a523cb","Type":"ContainerDied","Data":"666d32da5e6c45a1d48d0ef0c76ed76b0c47ad73e5b8a6c0440d75e7ee7b6d64"} Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713922 5018 scope.go:117] "RemoveContainer" containerID="5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.713932 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.745954 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.752050 5018 scope.go:117] "RemoveContainer" containerID="9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.772036 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.784170 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:06 crc kubenswrapper[5018]: E1014 08:48:06.784693 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.784711 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api" Oct 14 08:48:06 crc kubenswrapper[5018]: E1014 08:48:06.784723 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api-log" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.784731 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api-log" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.785023 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.785043 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" containerName="cinder-api-log" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.786260 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.791313 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.791831 5018 scope.go:117] "RemoveContainer" containerID="5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792190 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792229 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9lbdk" Oct 14 08:48:06 crc kubenswrapper[5018]: E1014 08:48:06.792417 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e\": container with ID starting with 5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e not found: ID does not exist" containerID="5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792453 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e"} err="failed to get container status \"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e\": rpc error: code = NotFound desc = could not find container \"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e\": container with ID starting with 5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e not found: ID does not exist" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792479 5018 scope.go:117] "RemoveContainer" containerID="9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792508 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.792807 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 08:48:06 crc kubenswrapper[5018]: E1014 08:48:06.792985 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1\": container with ID starting with 9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1 not found: ID does not exist" containerID="9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.793019 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.793031 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1"} err="failed to get container status \"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1\": rpc error: code = NotFound desc = could not find container \"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1\": container with ID starting with 9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1 not found: ID does not exist" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.793065 5018 scope.go:117] "RemoveContainer" containerID="5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.796865 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e"} err="failed to get container status \"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e\": rpc error: code = NotFound desc = could not find container \"5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e\": container with ID starting with 5a2b3afd8640f5a41b961236ee67d611b13359df09ffbb8d4693445c10c1ec1e not found: ID does not exist" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.796905 5018 scope.go:117] "RemoveContainer" containerID="9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.797268 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1"} err="failed to get container status \"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1\": rpc error: code = NotFound desc = could not find container \"9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1\": container with ID starting with 9a3919a216a1134496c46ea61c58e74abfc605839c164795cfb7d1e79f2e37a1 not found: ID does not exist" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.807352 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.902381 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.902837 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.902902 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.902986 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.903232 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.903290 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.903318 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.903383 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:06 crc kubenswrapper[5018]: I1014 08:48:06.903806 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.005313 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.005376 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.005460 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.005670 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006102 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006149 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006177 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006206 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006470 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.006913 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.009384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.010259 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.010853 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.014446 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.016032 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.018330 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.028647 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c\") pod \"cinder-api-0\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.113786 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.696223 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:07 crc kubenswrapper[5018]: I1014 08:48:07.735028 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerStarted","Data":"28fac02a8b3667b32916e3e3628930bf1d85963d6a9645b0b5440bc0d568df5f"} Oct 14 08:48:08 crc kubenswrapper[5018]: I1014 08:48:08.534286 5018 scope.go:117] "RemoveContainer" containerID="6ec0fca037768af2c6ab54de01240b5c1ec59460aa82d49fee10b48b4e0cd173" Oct 14 08:48:08 crc kubenswrapper[5018]: I1014 08:48:08.579904 5018 scope.go:117] "RemoveContainer" containerID="ad988727da44fbb625c5d40108ba5ec483d18f15f3181e6860c2a9f5b2ae2707" Oct 14 08:48:08 crc kubenswrapper[5018]: I1014 08:48:08.629660 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1c1369-34a3-48a5-acfb-bb4232a523cb" path="/var/lib/kubelet/pods/2d1c1369-34a3-48a5-acfb-bb4232a523cb/volumes" Oct 14 08:48:08 crc kubenswrapper[5018]: I1014 08:48:08.753151 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerStarted","Data":"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799"} Oct 14 08:48:09 crc kubenswrapper[5018]: I1014 08:48:09.769615 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerStarted","Data":"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812"} Oct 14 08:48:09 crc kubenswrapper[5018]: I1014 08:48:09.770780 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 08:48:09 crc kubenswrapper[5018]: I1014 08:48:09.818079 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.818051318 podStartE2EDuration="3.818051318s" podCreationTimestamp="2025-10-14 08:48:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:48:09.79965708 +0000 UTC m=+7106.383703747" watchObservedRunningTime="2025-10-14 08:48:09.818051318 +0000 UTC m=+7106.402097975" Oct 14 08:48:10 crc kubenswrapper[5018]: I1014 08:48:10.606241 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:48:10 crc kubenswrapper[5018]: E1014 08:48:10.607266 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:48:12 crc kubenswrapper[5018]: I1014 08:48:12.756949 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:48:12 crc kubenswrapper[5018]: I1014 08:48:12.869418 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:48:12 crc kubenswrapper[5018]: I1014 08:48:12.869688 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="dnsmasq-dns" containerID="cri-o://861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb" gracePeriod=10 Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.359770 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.535125 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b25rb\" (UniqueName: \"kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb\") pod \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.535240 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb\") pod \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.535291 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc\") pod \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.535336 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config\") pod \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.535398 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb\") pod \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\" (UID: \"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b\") " Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.545174 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb" (OuterVolumeSpecName: "kube-api-access-b25rb") pod "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" (UID: "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b"). InnerVolumeSpecName "kube-api-access-b25rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.585388 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config" (OuterVolumeSpecName: "config") pod "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" (UID: "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.586958 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" (UID: "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.592477 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" (UID: "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.598591 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" (UID: "4f7c960f-3c71-4fe3-b5f2-c5fe7296464b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.636943 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.636978 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.636986 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.636994 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.637004 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b25rb\" (UniqueName: \"kubernetes.io/projected/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b-kube-api-access-b25rb\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.814589 5018 generic.go:334] "Generic (PLEG): container finished" podID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerID="861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb" exitCode=0 Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.814709 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.814672 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" event={"ID":"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b","Type":"ContainerDied","Data":"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb"} Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.814878 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8569859d79-xs6xq" event={"ID":"4f7c960f-3c71-4fe3-b5f2-c5fe7296464b","Type":"ContainerDied","Data":"07674e4671014f35ecf72a6cb1338cfe65072695af02d0390b1793aff3ec495c"} Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.814918 5018 scope.go:117] "RemoveContainer" containerID="861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.846461 5018 scope.go:117] "RemoveContainer" containerID="9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.872303 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.884862 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8569859d79-xs6xq"] Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.885292 5018 scope.go:117] "RemoveContainer" containerID="861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb" Oct 14 08:48:13 crc kubenswrapper[5018]: E1014 08:48:13.885831 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb\": container with ID starting with 861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb not found: ID does not exist" containerID="861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.885896 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb"} err="failed to get container status \"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb\": rpc error: code = NotFound desc = could not find container \"861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb\": container with ID starting with 861d09d4bfc7b9b51dedd1f2d0c68b6ad06a8444ae87b5af4304c03e545e42eb not found: ID does not exist" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.885931 5018 scope.go:117] "RemoveContainer" containerID="9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2" Oct 14 08:48:13 crc kubenswrapper[5018]: E1014 08:48:13.886320 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2\": container with ID starting with 9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2 not found: ID does not exist" containerID="9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2" Oct 14 08:48:13 crc kubenswrapper[5018]: I1014 08:48:13.886388 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2"} err="failed to get container status \"9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2\": rpc error: code = NotFound desc = could not find container \"9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2\": container with ID starting with 9670519f9b87b9a8288c18f895d4fd50f082758c48569a94263f235fb86fd0b2 not found: ID does not exist" Oct 14 08:48:14 crc kubenswrapper[5018]: I1014 08:48:14.624970 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" path="/var/lib/kubelet/pods/4f7c960f-3c71-4fe3-b5f2-c5fe7296464b/volumes" Oct 14 08:48:18 crc kubenswrapper[5018]: I1014 08:48:18.799911 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 08:48:23 crc kubenswrapper[5018]: I1014 08:48:23.604672 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:48:23 crc kubenswrapper[5018]: E1014 08:48:23.605435 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:48:29 crc kubenswrapper[5018]: E1014 08:48:29.966667 5018 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.159:40070->38.102.83.159:32939: write tcp 38.102.83.159:40070->38.102.83.159:32939: write: broken pipe Oct 14 08:48:35 crc kubenswrapper[5018]: I1014 08:48:35.605090 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:48:35 crc kubenswrapper[5018]: E1014 08:48:35.608135 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.712668 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:40 crc kubenswrapper[5018]: E1014 08:48:40.713866 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="init" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.713880 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="init" Oct 14 08:48:40 crc kubenswrapper[5018]: E1014 08:48:40.713908 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="dnsmasq-dns" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.713914 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="dnsmasq-dns" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.714212 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7c960f-3c71-4fe3-b5f2-c5fe7296464b" containerName="dnsmasq-dns" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.715561 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.737935 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.741056 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924662 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924827 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b22t\" (UniqueName: \"kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924864 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924885 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924908 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:40 crc kubenswrapper[5018]: I1014 08:48:40.924931 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026239 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026355 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b22t\" (UniqueName: \"kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026383 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026400 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026416 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026431 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.026547 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.035340 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.035576 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.035898 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.036297 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.048289 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b22t\" (UniqueName: \"kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t\") pod \"cinder-scheduler-0\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.057078 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:41 crc kubenswrapper[5018]: I1014 08:48:41.670941 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:42 crc kubenswrapper[5018]: I1014 08:48:42.115916 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerStarted","Data":"1fde7ef995b41c7d5f9b36b0baacca57f6349c7012df4dc711239c093c8d6fd9"} Oct 14 08:48:42 crc kubenswrapper[5018]: I1014 08:48:42.129417 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:42 crc kubenswrapper[5018]: I1014 08:48:42.129684 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api-log" containerID="cri-o://644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799" gracePeriod=30 Oct 14 08:48:42 crc kubenswrapper[5018]: I1014 08:48:42.129769 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api" containerID="cri-o://0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812" gracePeriod=30 Oct 14 08:48:43 crc kubenswrapper[5018]: I1014 08:48:43.127461 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerStarted","Data":"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674"} Oct 14 08:48:43 crc kubenswrapper[5018]: I1014 08:48:43.131447 5018 generic.go:334] "Generic (PLEG): container finished" podID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerID="644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799" exitCode=143 Oct 14 08:48:43 crc kubenswrapper[5018]: I1014 08:48:43.131483 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerDied","Data":"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799"} Oct 14 08:48:44 crc kubenswrapper[5018]: I1014 08:48:44.161932 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerStarted","Data":"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c"} Oct 14 08:48:44 crc kubenswrapper[5018]: I1014 08:48:44.194186 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.824307087 podStartE2EDuration="4.194154774s" podCreationTimestamp="2025-10-14 08:48:40 +0000 UTC" firstStartedPulling="2025-10-14 08:48:41.677224312 +0000 UTC m=+7138.261270939" lastFinishedPulling="2025-10-14 08:48:42.047071989 +0000 UTC m=+7138.631118626" observedRunningTime="2025-10-14 08:48:44.192756934 +0000 UTC m=+7140.776803611" watchObservedRunningTime="2025-10-14 08:48:44.194154774 +0000 UTC m=+7140.778201441" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.689988 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781001 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781154 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781381 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781573 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781682 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781754 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.781857 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.782089 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.782131 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.782749 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs\") pod \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\" (UID: \"a379bdb0-1f30-45c4-b6dd-2c93144d66b7\") " Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.782614 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs" (OuterVolumeSpecName: "logs") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.783390 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.783416 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.789692 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.789894 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c" (OuterVolumeSpecName: "kube-api-access-gj22c") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "kube-api-access-gj22c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.791208 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts" (OuterVolumeSpecName: "scripts") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.842066 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.864598 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.867646 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data" (OuterVolumeSpecName: "config-data") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.872856 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a379bdb0-1f30-45c4-b6dd-2c93144d66b7" (UID: "a379bdb0-1f30-45c4-b6dd-2c93144d66b7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884504 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj22c\" (UniqueName: \"kubernetes.io/projected/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-kube-api-access-gj22c\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884543 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884552 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884561 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884571 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884579 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:45 crc kubenswrapper[5018]: I1014 08:48:45.884588 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a379bdb0-1f30-45c4-b6dd-2c93144d66b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.058514 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.196065 5018 generic.go:334] "Generic (PLEG): container finished" podID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerID="0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812" exitCode=0 Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.196117 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerDied","Data":"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812"} Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.196214 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a379bdb0-1f30-45c4-b6dd-2c93144d66b7","Type":"ContainerDied","Data":"28fac02a8b3667b32916e3e3628930bf1d85963d6a9645b0b5440bc0d568df5f"} Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.196139 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.196247 5018 scope.go:117] "RemoveContainer" containerID="0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.241294 5018 scope.go:117] "RemoveContainer" containerID="644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.241837 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.265170 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.316806 5018 scope.go:117] "RemoveContainer" containerID="0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.316815 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:46 crc kubenswrapper[5018]: E1014 08:48:46.317497 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api-log" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.317554 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api-log" Oct 14 08:48:46 crc kubenswrapper[5018]: E1014 08:48:46.317577 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.317587 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.318014 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.318081 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" containerName="cinder-api-log" Oct 14 08:48:46 crc kubenswrapper[5018]: E1014 08:48:46.318743 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812\": container with ID starting with 0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812 not found: ID does not exist" containerID="0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.318772 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812"} err="failed to get container status \"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812\": rpc error: code = NotFound desc = could not find container \"0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812\": container with ID starting with 0c48ed3c10ecf861d1ca979bd09495ba054a99483ae402ca3677485c9fd10812 not found: ID does not exist" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.318793 5018 scope.go:117] "RemoveContainer" containerID="644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799" Oct 14 08:48:46 crc kubenswrapper[5018]: E1014 08:48:46.319128 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799\": container with ID starting with 644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799 not found: ID does not exist" containerID="644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.319152 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799"} err="failed to get container status \"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799\": rpc error: code = NotFound desc = could not find container \"644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799\": container with ID starting with 644c1e74e76477d0a6a663b129105c2a23a9d673ea29fa2b142b5890e1e14799 not found: ID does not exist" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.320116 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.322083 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.322666 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.323327 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.323555 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.417872 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418083 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crpr6\" (UniqueName: \"kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418117 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418148 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418189 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418212 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418463 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418496 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.418520 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520517 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520649 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crpr6\" (UniqueName: \"kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520683 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520720 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520772 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520838 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.520954 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.521006 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.521029 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.521078 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.522607 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.524255 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.524871 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.525913 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.526612 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.527070 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.531140 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.537546 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crpr6\" (UniqueName: \"kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6\") pod \"cinder-api-0\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " pod="openstack/cinder-api-0" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.605438 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:48:46 crc kubenswrapper[5018]: E1014 08:48:46.606368 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.618350 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a379bdb0-1f30-45c4-b6dd-2c93144d66b7" path="/var/lib/kubelet/pods/a379bdb0-1f30-45c4-b6dd-2c93144d66b7/volumes" Oct 14 08:48:46 crc kubenswrapper[5018]: I1014 08:48:46.650678 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:48:47 crc kubenswrapper[5018]: I1014 08:48:47.146971 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:48:47 crc kubenswrapper[5018]: W1014 08:48:47.157521 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07df2031_64c4_4e6b_b70a_831edefc4468.slice/crio-4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec WatchSource:0}: Error finding container 4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec: Status 404 returned error can't find the container with id 4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec Oct 14 08:48:47 crc kubenswrapper[5018]: I1014 08:48:47.210459 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerStarted","Data":"4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec"} Oct 14 08:48:48 crc kubenswrapper[5018]: I1014 08:48:48.226914 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerStarted","Data":"ce612af3722b887703ef66d47d90f1bdbe0b44f623bc74647fe5ba7231238eed"} Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.247311 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerStarted","Data":"cc378e4ea02d43f477ec8203ef137c674e7055a3b57a90d0bac877f9fdd800da"} Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.247783 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.284804 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.284776589 podStartE2EDuration="3.284776589s" podCreationTimestamp="2025-10-14 08:48:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:48:49.273003081 +0000 UTC m=+7145.857049748" watchObservedRunningTime="2025-10-14 08:48:49.284776589 +0000 UTC m=+7145.868823256" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.677436 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.681086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.688537 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.786517 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8fdh\" (UniqueName: \"kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.786701 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.786968 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.888386 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8fdh\" (UniqueName: \"kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.888457 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.888568 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.889235 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.889250 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:49 crc kubenswrapper[5018]: I1014 08:48:49.911984 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8fdh\" (UniqueName: \"kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh\") pod \"redhat-operators-4h74j\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:50 crc kubenswrapper[5018]: I1014 08:48:50.008713 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:48:50 crc kubenswrapper[5018]: I1014 08:48:50.534439 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.251715 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.270767 5018 generic.go:334] "Generic (PLEG): container finished" podID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerID="e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc" exitCode=0 Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.270823 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerDied","Data":"e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc"} Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.270855 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerStarted","Data":"51955c1f4b944e0f8ea056edf54060d7830f825e1ef729f0fb81fd5bc28364c4"} Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.314603 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.314879 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="cinder-scheduler" containerID="cri-o://aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674" gracePeriod=30 Oct 14 08:48:51 crc kubenswrapper[5018]: I1014 08:48:51.315026 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="probe" containerID="cri-o://7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c" gracePeriod=30 Oct 14 08:48:52 crc kubenswrapper[5018]: I1014 08:48:52.304146 5018 generic.go:334] "Generic (PLEG): container finished" podID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerID="7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c" exitCode=0 Oct 14 08:48:52 crc kubenswrapper[5018]: I1014 08:48:52.304455 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerDied","Data":"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c"} Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.319472 5018 generic.go:334] "Generic (PLEG): container finished" podID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerID="966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95" exitCode=0 Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.319876 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerDied","Data":"966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95"} Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.717027 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884411 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b22t\" (UniqueName: \"kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884509 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884534 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884572 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884649 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.884779 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle\") pod \"47db55cb-f7d0-4fa8-95ec-471629cf1360\" (UID: \"47db55cb-f7d0-4fa8-95ec-471629cf1360\") " Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.885913 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.891281 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.891313 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts" (OuterVolumeSpecName: "scripts") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.893813 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t" (OuterVolumeSpecName: "kube-api-access-7b22t") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "kube-api-access-7b22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.953395 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.987128 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.987320 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.987405 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47db55cb-f7d0-4fa8-95ec-471629cf1360-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.987491 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:53 crc kubenswrapper[5018]: I1014 08:48:53.987570 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b22t\" (UniqueName: \"kubernetes.io/projected/47db55cb-f7d0-4fa8-95ec-471629cf1360-kube-api-access-7b22t\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.003678 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data" (OuterVolumeSpecName: "config-data") pod "47db55cb-f7d0-4fa8-95ec-471629cf1360" (UID: "47db55cb-f7d0-4fa8-95ec-471629cf1360"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.089274 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47db55cb-f7d0-4fa8-95ec-471629cf1360-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.333147 5018 generic.go:334] "Generic (PLEG): container finished" podID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerID="aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674" exitCode=0 Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.333215 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.333212 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerDied","Data":"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674"} Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.333328 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"47db55cb-f7d0-4fa8-95ec-471629cf1360","Type":"ContainerDied","Data":"1fde7ef995b41c7d5f9b36b0baacca57f6349c7012df4dc711239c093c8d6fd9"} Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.333360 5018 scope.go:117] "RemoveContainer" containerID="7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.339889 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerStarted","Data":"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff"} Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.366703 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4h74j" podStartSLOduration=2.813994623 podStartE2EDuration="5.366676692s" podCreationTimestamp="2025-10-14 08:48:49 +0000 UTC" firstStartedPulling="2025-10-14 08:48:51.27253604 +0000 UTC m=+7147.856582667" lastFinishedPulling="2025-10-14 08:48:53.825218109 +0000 UTC m=+7150.409264736" observedRunningTime="2025-10-14 08:48:54.361386721 +0000 UTC m=+7150.945433348" watchObservedRunningTime="2025-10-14 08:48:54.366676692 +0000 UTC m=+7150.950723339" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.409542 5018 scope.go:117] "RemoveContainer" containerID="aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.409804 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.423760 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.436291 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:54 crc kubenswrapper[5018]: E1014 08:48:54.436820 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="cinder-scheduler" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.436843 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="cinder-scheduler" Oct 14 08:48:54 crc kubenswrapper[5018]: E1014 08:48:54.436871 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="probe" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.436880 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="probe" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.437082 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="cinder-scheduler" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.437115 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" containerName="probe" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.438253 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.442190 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.460283 5018 scope.go:117] "RemoveContainer" containerID="7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c" Oct 14 08:48:54 crc kubenswrapper[5018]: E1014 08:48:54.462333 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c\": container with ID starting with 7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c not found: ID does not exist" containerID="7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.462388 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c"} err="failed to get container status \"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c\": rpc error: code = NotFound desc = could not find container \"7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c\": container with ID starting with 7df3003a6b0a2b37db8b5beb316a930c14f42d72b6f4ed6efe7f0dfac3331b5c not found: ID does not exist" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.462423 5018 scope.go:117] "RemoveContainer" containerID="aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.462532 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:54 crc kubenswrapper[5018]: E1014 08:48:54.464320 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674\": container with ID starting with aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674 not found: ID does not exist" containerID="aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.464357 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674"} err="failed to get container status \"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674\": rpc error: code = NotFound desc = could not find container \"aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674\": container with ID starting with aaed1d334a8d675a0065a1708565f630168370cb3de04808f30ee6881e37b674 not found: ID does not exist" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.598758 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.598911 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.598982 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.599109 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.599230 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqqt2\" (UniqueName: \"kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.599317 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.619334 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47db55cb-f7d0-4fa8-95ec-471629cf1360" path="/var/lib/kubelet/pods/47db55cb-f7d0-4fa8-95ec-471629cf1360/volumes" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701128 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701224 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqqt2\" (UniqueName: \"kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701265 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701364 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701412 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701447 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.701941 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.706345 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.707937 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.711489 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.711764 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.721398 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqqt2\" (UniqueName: \"kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2\") pod \"cinder-scheduler-0\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " pod="openstack/cinder-scheduler-0" Oct 14 08:48:54 crc kubenswrapper[5018]: I1014 08:48:54.769970 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:48:55 crc kubenswrapper[5018]: I1014 08:48:55.311141 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:48:55 crc kubenswrapper[5018]: W1014 08:48:55.315175 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b5c5d38_56a1_447d_96f4_a8427ff3df82.slice/crio-bf7cca211ae8175bfa6033459b6e1fe68996cb35b76e89f90d6110438a73cc63 WatchSource:0}: Error finding container bf7cca211ae8175bfa6033459b6e1fe68996cb35b76e89f90d6110438a73cc63: Status 404 returned error can't find the container with id bf7cca211ae8175bfa6033459b6e1fe68996cb35b76e89f90d6110438a73cc63 Oct 14 08:48:55 crc kubenswrapper[5018]: I1014 08:48:55.374745 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerStarted","Data":"bf7cca211ae8175bfa6033459b6e1fe68996cb35b76e89f90d6110438a73cc63"} Oct 14 08:48:56 crc kubenswrapper[5018]: I1014 08:48:56.419148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerStarted","Data":"3c0372fbfcb7d77786b28550cb757bcf2d36775c128fc8d61c2eeed2bb4a27a4"} Oct 14 08:48:57 crc kubenswrapper[5018]: I1014 08:48:57.432230 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerStarted","Data":"ac50df93bbc18e55844ca557b4aa55078662999d23f67f0d58048a5477b2f20c"} Oct 14 08:48:57 crc kubenswrapper[5018]: I1014 08:48:57.470969 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.470943194 podStartE2EDuration="3.470943194s" podCreationTimestamp="2025-10-14 08:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:48:57.459892907 +0000 UTC m=+7154.043939534" watchObservedRunningTime="2025-10-14 08:48:57.470943194 +0000 UTC m=+7154.054989861" Oct 14 08:48:57 crc kubenswrapper[5018]: I1014 08:48:57.604670 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:48:57 crc kubenswrapper[5018]: E1014 08:48:57.604914 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:48:58 crc kubenswrapper[5018]: I1014 08:48:58.479281 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 08:48:59 crc kubenswrapper[5018]: I1014 08:48:59.770699 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 08:49:00 crc kubenswrapper[5018]: I1014 08:49:00.009025 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:00 crc kubenswrapper[5018]: I1014 08:49:00.009116 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:00 crc kubenswrapper[5018]: I1014 08:49:00.083781 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:00 crc kubenswrapper[5018]: I1014 08:49:00.548165 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:00 crc kubenswrapper[5018]: I1014 08:49:00.626243 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:49:02 crc kubenswrapper[5018]: I1014 08:49:02.492729 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4h74j" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="registry-server" containerID="cri-o://f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff" gracePeriod=2 Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.004576 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.067089 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8fdh\" (UniqueName: \"kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh\") pod \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.067147 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content\") pod \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.067199 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities\") pod \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\" (UID: \"8f591c8f-f5d3-49e8-968d-2d167bedc42c\") " Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.067928 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities" (OuterVolumeSpecName: "utilities") pod "8f591c8f-f5d3-49e8-968d-2d167bedc42c" (UID: "8f591c8f-f5d3-49e8-968d-2d167bedc42c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.068805 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.079257 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh" (OuterVolumeSpecName: "kube-api-access-x8fdh") pod "8f591c8f-f5d3-49e8-968d-2d167bedc42c" (UID: "8f591c8f-f5d3-49e8-968d-2d167bedc42c"). InnerVolumeSpecName "kube-api-access-x8fdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.156022 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f591c8f-f5d3-49e8-968d-2d167bedc42c" (UID: "8f591c8f-f5d3-49e8-968d-2d167bedc42c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.171236 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8fdh\" (UniqueName: \"kubernetes.io/projected/8f591c8f-f5d3-49e8-968d-2d167bedc42c-kube-api-access-x8fdh\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.171293 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f591c8f-f5d3-49e8-968d-2d167bedc42c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.505491 5018 generic.go:334] "Generic (PLEG): container finished" podID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerID="f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff" exitCode=0 Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.505589 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4h74j" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.505588 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerDied","Data":"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff"} Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.505968 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4h74j" event={"ID":"8f591c8f-f5d3-49e8-968d-2d167bedc42c","Type":"ContainerDied","Data":"51955c1f4b944e0f8ea056edf54060d7830f825e1ef729f0fb81fd5bc28364c4"} Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.506048 5018 scope.go:117] "RemoveContainer" containerID="f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.536623 5018 scope.go:117] "RemoveContainer" containerID="966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.558704 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.566850 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4h74j"] Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.592483 5018 scope.go:117] "RemoveContainer" containerID="e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.616432 5018 scope.go:117] "RemoveContainer" containerID="f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff" Oct 14 08:49:03 crc kubenswrapper[5018]: E1014 08:49:03.616900 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff\": container with ID starting with f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff not found: ID does not exist" containerID="f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.616956 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff"} err="failed to get container status \"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff\": rpc error: code = NotFound desc = could not find container \"f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff\": container with ID starting with f5d4804b4bb04848793a27760a02d28b52986066a43ff4ead840235b9f6032ff not found: ID does not exist" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.616977 5018 scope.go:117] "RemoveContainer" containerID="966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95" Oct 14 08:49:03 crc kubenswrapper[5018]: E1014 08:49:03.617301 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95\": container with ID starting with 966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95 not found: ID does not exist" containerID="966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.617354 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95"} err="failed to get container status \"966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95\": rpc error: code = NotFound desc = could not find container \"966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95\": container with ID starting with 966ff1be0b39f88ea74f41fe24f2b96cdb87b1f39e3b588934aab2edccc1df95 not found: ID does not exist" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.617370 5018 scope.go:117] "RemoveContainer" containerID="e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc" Oct 14 08:49:03 crc kubenswrapper[5018]: E1014 08:49:03.617718 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc\": container with ID starting with e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc not found: ID does not exist" containerID="e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc" Oct 14 08:49:03 crc kubenswrapper[5018]: I1014 08:49:03.617766 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc"} err="failed to get container status \"e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc\": rpc error: code = NotFound desc = could not find container \"e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc\": container with ID starting with e1f8c7fa9f03b9496161c5aee8a5f454b5ae8e306865b9a9ed1cf699ef1b72dc not found: ID does not exist" Oct 14 08:49:04 crc kubenswrapper[5018]: I1014 08:49:04.645268 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" path="/var/lib/kubelet/pods/8f591c8f-f5d3-49e8-968d-2d167bedc42c/volumes" Oct 14 08:49:05 crc kubenswrapper[5018]: I1014 08:49:05.081107 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.707190 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ttc4f"] Oct 14 08:49:07 crc kubenswrapper[5018]: E1014 08:49:07.709543 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="extract-content" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.709568 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="extract-content" Oct 14 08:49:07 crc kubenswrapper[5018]: E1014 08:49:07.709593 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="extract-utilities" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.709607 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="extract-utilities" Oct 14 08:49:07 crc kubenswrapper[5018]: E1014 08:49:07.709674 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="registry-server" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.709687 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="registry-server" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.709969 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f591c8f-f5d3-49e8-968d-2d167bedc42c" containerName="registry-server" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.711037 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.715769 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ttc4f"] Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.885099 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6dtb\" (UniqueName: \"kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb\") pod \"glance-db-create-ttc4f\" (UID: \"9d53c1c9-9ebf-4356-b632-7152d87b8055\") " pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:07 crc kubenswrapper[5018]: I1014 08:49:07.986886 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6dtb\" (UniqueName: \"kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb\") pod \"glance-db-create-ttc4f\" (UID: \"9d53c1c9-9ebf-4356-b632-7152d87b8055\") " pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:08 crc kubenswrapper[5018]: I1014 08:49:08.019089 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6dtb\" (UniqueName: \"kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb\") pod \"glance-db-create-ttc4f\" (UID: \"9d53c1c9-9ebf-4356-b632-7152d87b8055\") " pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:08 crc kubenswrapper[5018]: I1014 08:49:08.038583 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:08 crc kubenswrapper[5018]: I1014 08:49:08.542726 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ttc4f"] Oct 14 08:49:08 crc kubenswrapper[5018]: W1014 08:49:08.545414 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d53c1c9_9ebf_4356_b632_7152d87b8055.slice/crio-5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281 WatchSource:0}: Error finding container 5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281: Status 404 returned error can't find the container with id 5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281 Oct 14 08:49:08 crc kubenswrapper[5018]: I1014 08:49:08.566964 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ttc4f" event={"ID":"9d53c1c9-9ebf-4356-b632-7152d87b8055","Type":"ContainerStarted","Data":"5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281"} Oct 14 08:49:08 crc kubenswrapper[5018]: I1014 08:49:08.702831 5018 scope.go:117] "RemoveContainer" containerID="dcd4c395ebfeabaaa50fe723fd4991e8d8c4491d321f34937e11bdb35db53078" Oct 14 08:49:09 crc kubenswrapper[5018]: I1014 08:49:09.580541 5018 generic.go:334] "Generic (PLEG): container finished" podID="9d53c1c9-9ebf-4356-b632-7152d87b8055" containerID="3e0af616afa486a9141c3753b888e9b5377cd72aed3a6a0eccbee9d8d201c317" exitCode=0 Oct 14 08:49:09 crc kubenswrapper[5018]: I1014 08:49:09.580668 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ttc4f" event={"ID":"9d53c1c9-9ebf-4356-b632-7152d87b8055","Type":"ContainerDied","Data":"3e0af616afa486a9141c3753b888e9b5377cd72aed3a6a0eccbee9d8d201c317"} Oct 14 08:49:10 crc kubenswrapper[5018]: I1014 08:49:10.975264 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.146427 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6dtb\" (UniqueName: \"kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb\") pod \"9d53c1c9-9ebf-4356-b632-7152d87b8055\" (UID: \"9d53c1c9-9ebf-4356-b632-7152d87b8055\") " Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.154532 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb" (OuterVolumeSpecName: "kube-api-access-f6dtb") pod "9d53c1c9-9ebf-4356-b632-7152d87b8055" (UID: "9d53c1c9-9ebf-4356-b632-7152d87b8055"). InnerVolumeSpecName "kube-api-access-f6dtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.248330 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6dtb\" (UniqueName: \"kubernetes.io/projected/9d53c1c9-9ebf-4356-b632-7152d87b8055-kube-api-access-f6dtb\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.604850 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ttc4f" event={"ID":"9d53c1c9-9ebf-4356-b632-7152d87b8055","Type":"ContainerDied","Data":"5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281"} Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.604911 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aac35b8d02b6cd7a71dbe08d10baced772a08722c7175dcd3530166a4e70281" Oct 14 08:49:11 crc kubenswrapper[5018]: I1014 08:49:11.605113 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ttc4f" Oct 14 08:49:12 crc kubenswrapper[5018]: I1014 08:49:12.605051 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:49:12 crc kubenswrapper[5018]: E1014 08:49:12.605785 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.864337 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-72c6-account-create-5hkcx"] Oct 14 08:49:17 crc kubenswrapper[5018]: E1014 08:49:17.865598 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d53c1c9-9ebf-4356-b632-7152d87b8055" containerName="mariadb-database-create" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.865771 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d53c1c9-9ebf-4356-b632-7152d87b8055" containerName="mariadb-database-create" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.866227 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d53c1c9-9ebf-4356-b632-7152d87b8055" containerName="mariadb-database-create" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.867206 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.869515 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 08:49:17 crc kubenswrapper[5018]: I1014 08:49:17.872980 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-72c6-account-create-5hkcx"] Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.009879 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdrs\" (UniqueName: \"kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs\") pod \"glance-72c6-account-create-5hkcx\" (UID: \"45fc70f1-8084-4a78-bbd1-0ee01009bdd8\") " pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.112928 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cdrs\" (UniqueName: \"kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs\") pod \"glance-72c6-account-create-5hkcx\" (UID: \"45fc70f1-8084-4a78-bbd1-0ee01009bdd8\") " pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.139524 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cdrs\" (UniqueName: \"kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs\") pod \"glance-72c6-account-create-5hkcx\" (UID: \"45fc70f1-8084-4a78-bbd1-0ee01009bdd8\") " pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.201327 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.510215 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-72c6-account-create-5hkcx"] Oct 14 08:49:18 crc kubenswrapper[5018]: W1014 08:49:18.522469 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45fc70f1_8084_4a78_bbd1_0ee01009bdd8.slice/crio-0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3 WatchSource:0}: Error finding container 0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3: Status 404 returned error can't find the container with id 0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3 Oct 14 08:49:18 crc kubenswrapper[5018]: I1014 08:49:18.677900 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-72c6-account-create-5hkcx" event={"ID":"45fc70f1-8084-4a78-bbd1-0ee01009bdd8","Type":"ContainerStarted","Data":"0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3"} Oct 14 08:49:18 crc kubenswrapper[5018]: E1014 08:49:18.942053 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45fc70f1_8084_4a78_bbd1_0ee01009bdd8.slice/crio-62b082682f42b865dbdfcdf0a9e426152caf523ebbdc379a94547fffe07b3870.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45fc70f1_8084_4a78_bbd1_0ee01009bdd8.slice/crio-conmon-62b082682f42b865dbdfcdf0a9e426152caf523ebbdc379a94547fffe07b3870.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:49:19 crc kubenswrapper[5018]: I1014 08:49:19.695048 5018 generic.go:334] "Generic (PLEG): container finished" podID="45fc70f1-8084-4a78-bbd1-0ee01009bdd8" containerID="62b082682f42b865dbdfcdf0a9e426152caf523ebbdc379a94547fffe07b3870" exitCode=0 Oct 14 08:49:19 crc kubenswrapper[5018]: I1014 08:49:19.695127 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-72c6-account-create-5hkcx" event={"ID":"45fc70f1-8084-4a78-bbd1-0ee01009bdd8","Type":"ContainerDied","Data":"62b082682f42b865dbdfcdf0a9e426152caf523ebbdc379a94547fffe07b3870"} Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.169564 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.277847 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cdrs\" (UniqueName: \"kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs\") pod \"45fc70f1-8084-4a78-bbd1-0ee01009bdd8\" (UID: \"45fc70f1-8084-4a78-bbd1-0ee01009bdd8\") " Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.287500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs" (OuterVolumeSpecName: "kube-api-access-4cdrs") pod "45fc70f1-8084-4a78-bbd1-0ee01009bdd8" (UID: "45fc70f1-8084-4a78-bbd1-0ee01009bdd8"). InnerVolumeSpecName "kube-api-access-4cdrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.380566 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cdrs\" (UniqueName: \"kubernetes.io/projected/45fc70f1-8084-4a78-bbd1-0ee01009bdd8-kube-api-access-4cdrs\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.719471 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-72c6-account-create-5hkcx" event={"ID":"45fc70f1-8084-4a78-bbd1-0ee01009bdd8","Type":"ContainerDied","Data":"0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3"} Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.719510 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0347c49625b1e62b95ccb845871dc134d5e3d42872ca2d551b87c4a8a75a03f3" Oct 14 08:49:21 crc kubenswrapper[5018]: I1014 08:49:21.719550 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-72c6-account-create-5hkcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.031801 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-l4hcx"] Oct 14 08:49:23 crc kubenswrapper[5018]: E1014 08:49:23.032661 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fc70f1-8084-4a78-bbd1-0ee01009bdd8" containerName="mariadb-account-create" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.032681 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fc70f1-8084-4a78-bbd1-0ee01009bdd8" containerName="mariadb-account-create" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.032880 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="45fc70f1-8084-4a78-bbd1-0ee01009bdd8" containerName="mariadb-account-create" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.033590 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.035952 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zdjgk" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.035989 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.041757 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l4hcx"] Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.116860 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.116948 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.117120 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.117212 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xj7j\" (UniqueName: \"kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.219329 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.219404 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.219497 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.219539 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xj7j\" (UniqueName: \"kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.229463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.229463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.239412 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.242242 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xj7j\" (UniqueName: \"kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j\") pod \"glance-db-sync-l4hcx\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.372648 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.605450 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:49:23 crc kubenswrapper[5018]: E1014 08:49:23.607096 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:49:23 crc kubenswrapper[5018]: I1014 08:49:23.965286 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l4hcx"] Oct 14 08:49:24 crc kubenswrapper[5018]: I1014 08:49:24.743972 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l4hcx" event={"ID":"b57a52e6-9c24-4925-8503-c5e9cd43de14","Type":"ContainerStarted","Data":"24ca43e784e1d9c573ce5b6914a3429e750bcab9e617a56584f32eab643b5f49"} Oct 14 08:49:38 crc kubenswrapper[5018]: I1014 08:49:38.605013 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:49:38 crc kubenswrapper[5018]: E1014 08:49:38.606116 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:49:40 crc kubenswrapper[5018]: I1014 08:49:40.944214 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l4hcx" event={"ID":"b57a52e6-9c24-4925-8503-c5e9cd43de14","Type":"ContainerStarted","Data":"ce5d39a39c7ed7bbc9152f97a9bf47f809f0c78524ebd4d01a7d0eea008d0b17"} Oct 14 08:49:40 crc kubenswrapper[5018]: I1014 08:49:40.993114 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-l4hcx" podStartSLOduration=3.5190206059999998 podStartE2EDuration="18.993087232s" podCreationTimestamp="2025-10-14 08:49:22 +0000 UTC" firstStartedPulling="2025-10-14 08:49:23.978823792 +0000 UTC m=+7180.562870429" lastFinishedPulling="2025-10-14 08:49:39.452890428 +0000 UTC m=+7196.036937055" observedRunningTime="2025-10-14 08:49:40.981399377 +0000 UTC m=+7197.565446044" watchObservedRunningTime="2025-10-14 08:49:40.993087232 +0000 UTC m=+7197.577133929" Oct 14 08:49:43 crc kubenswrapper[5018]: I1014 08:49:43.981369 5018 generic.go:334] "Generic (PLEG): container finished" podID="b57a52e6-9c24-4925-8503-c5e9cd43de14" containerID="ce5d39a39c7ed7bbc9152f97a9bf47f809f0c78524ebd4d01a7d0eea008d0b17" exitCode=0 Oct 14 08:49:43 crc kubenswrapper[5018]: I1014 08:49:43.981421 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l4hcx" event={"ID":"b57a52e6-9c24-4925-8503-c5e9cd43de14","Type":"ContainerDied","Data":"ce5d39a39c7ed7bbc9152f97a9bf47f809f0c78524ebd4d01a7d0eea008d0b17"} Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.448774 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.559135 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle\") pod \"b57a52e6-9c24-4925-8503-c5e9cd43de14\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.559217 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data\") pod \"b57a52e6-9c24-4925-8503-c5e9cd43de14\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.559300 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data\") pod \"b57a52e6-9c24-4925-8503-c5e9cd43de14\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.559351 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xj7j\" (UniqueName: \"kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j\") pod \"b57a52e6-9c24-4925-8503-c5e9cd43de14\" (UID: \"b57a52e6-9c24-4925-8503-c5e9cd43de14\") " Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.566498 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b57a52e6-9c24-4925-8503-c5e9cd43de14" (UID: "b57a52e6-9c24-4925-8503-c5e9cd43de14"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.568579 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j" (OuterVolumeSpecName: "kube-api-access-5xj7j") pod "b57a52e6-9c24-4925-8503-c5e9cd43de14" (UID: "b57a52e6-9c24-4925-8503-c5e9cd43de14"). InnerVolumeSpecName "kube-api-access-5xj7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.605874 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b57a52e6-9c24-4925-8503-c5e9cd43de14" (UID: "b57a52e6-9c24-4925-8503-c5e9cd43de14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.655320 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data" (OuterVolumeSpecName: "config-data") pod "b57a52e6-9c24-4925-8503-c5e9cd43de14" (UID: "b57a52e6-9c24-4925-8503-c5e9cd43de14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.661987 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.662036 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.662056 5018 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b57a52e6-9c24-4925-8503-c5e9cd43de14-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:45 crc kubenswrapper[5018]: I1014 08:49:45.662074 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xj7j\" (UniqueName: \"kubernetes.io/projected/b57a52e6-9c24-4925-8503-c5e9cd43de14-kube-api-access-5xj7j\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.015460 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l4hcx" event={"ID":"b57a52e6-9c24-4925-8503-c5e9cd43de14","Type":"ContainerDied","Data":"24ca43e784e1d9c573ce5b6914a3429e750bcab9e617a56584f32eab643b5f49"} Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.015528 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24ca43e784e1d9c573ce5b6914a3429e750bcab9e617a56584f32eab643b5f49" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.015705 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l4hcx" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.393020 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:49:46 crc kubenswrapper[5018]: E1014 08:49:46.393758 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57a52e6-9c24-4925-8503-c5e9cd43de14" containerName="glance-db-sync" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.393776 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57a52e6-9c24-4925-8503-c5e9cd43de14" containerName="glance-db-sync" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.393945 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b57a52e6-9c24-4925-8503-c5e9cd43de14" containerName="glance-db-sync" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.394901 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.416209 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.417801 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.422114 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.422229 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zdjgk" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.422397 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.432743 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.441901 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.479328 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.481064 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.483466 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.489048 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579219 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579255 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579287 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579318 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579335 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579351 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579534 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579588 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579656 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgzf4\" (UniqueName: \"kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579683 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs7fb\" (UniqueName: \"kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579710 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579773 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579824 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579862 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgktg\" (UniqueName: \"kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579948 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.579979 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.580001 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.681991 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682040 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682081 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682123 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682147 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682169 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682209 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682232 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682262 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgzf4\" (UniqueName: \"kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682285 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs7fb\" (UniqueName: \"kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682316 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682356 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682396 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682424 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgktg\" (UniqueName: \"kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682471 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682496 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.682521 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683524 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683613 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683667 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683774 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683822 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.683991 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.684351 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.684428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.697325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.701951 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.702920 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.705994 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.706151 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.707193 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.712340 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgzf4\" (UniqueName: \"kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4\") pod \"glance-default-external-api-0\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.713642 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgktg\" (UniqueName: \"kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg\") pod \"dnsmasq-dns-5c664b6d67-8dmsz\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.714784 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs7fb\" (UniqueName: \"kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb\") pod \"glance-default-internal-api-0\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.760541 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.785041 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:46 crc kubenswrapper[5018]: I1014 08:49:46.806203 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:47 crc kubenswrapper[5018]: I1014 08:49:47.304753 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:49:47 crc kubenswrapper[5018]: I1014 08:49:47.373304 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:47 crc kubenswrapper[5018]: I1014 08:49:47.529164 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.044651 5018 generic.go:334] "Generic (PLEG): container finished" podID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerID="5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a" exitCode=0 Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.044706 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" event={"ID":"94719570-6817-4a41-8d9a-1ce46a4591e8","Type":"ContainerDied","Data":"5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a"} Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.044732 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" event={"ID":"94719570-6817-4a41-8d9a-1ce46a4591e8","Type":"ContainerStarted","Data":"826915519fecd109ddbdc4871a9a7ad2785aa1745111d6c6e884bb7be0fb7d79"} Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.047523 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerStarted","Data":"c674d15dfcf8da8bf7586e941a69d227db8465b246d3504d6fcfb5e811e6b5ff"} Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.477822 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:48 crc kubenswrapper[5018]: I1014 08:49:48.761957 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.056820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerStarted","Data":"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca"} Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.056862 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerStarted","Data":"e47251690103934223413685b99f1eae9ad15de8667464033a979884bca956a4"} Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.058371 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerStarted","Data":"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857"} Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.058394 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerStarted","Data":"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b"} Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.058517 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-log" containerID="cri-o://89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" gracePeriod=30 Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.059017 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-httpd" containerID="cri-o://cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" gracePeriod=30 Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.061123 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" event={"ID":"94719570-6817-4a41-8d9a-1ce46a4591e8","Type":"ContainerStarted","Data":"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628"} Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.061978 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.090296 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.090276273 podStartE2EDuration="3.090276273s" podCreationTimestamp="2025-10-14 08:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:49:49.084427285 +0000 UTC m=+7205.668473912" watchObservedRunningTime="2025-10-14 08:49:49.090276273 +0000 UTC m=+7205.674322900" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.110644 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" podStartSLOduration=3.110599767 podStartE2EDuration="3.110599767s" podCreationTimestamp="2025-10-14 08:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:49:49.107347623 +0000 UTC m=+7205.691394250" watchObservedRunningTime="2025-10-14 08:49:49.110599767 +0000 UTC m=+7205.694646394" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.707175 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735175 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735245 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735279 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735317 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgzf4\" (UniqueName: \"kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735387 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735440 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs\") pod \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\" (UID: \"ffc19213-da31-4ee6-b0db-5fa87acfdc45\") " Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.735736 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.736125 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs" (OuterVolumeSpecName: "logs") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.736260 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.736283 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ffc19213-da31-4ee6-b0db-5fa87acfdc45-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.759122 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4" (OuterVolumeSpecName: "kube-api-access-kgzf4") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "kube-api-access-kgzf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.762878 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts" (OuterVolumeSpecName: "scripts") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.776003 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.796635 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data" (OuterVolumeSpecName: "config-data") pod "ffc19213-da31-4ee6-b0db-5fa87acfdc45" (UID: "ffc19213-da31-4ee6-b0db-5fa87acfdc45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.837923 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.837961 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.837974 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgzf4\" (UniqueName: \"kubernetes.io/projected/ffc19213-da31-4ee6-b0db-5fa87acfdc45-kube-api-access-kgzf4\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:49 crc kubenswrapper[5018]: I1014 08:49:49.837985 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc19213-da31-4ee6-b0db-5fa87acfdc45-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.072850 5018 generic.go:334] "Generic (PLEG): container finished" podID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerID="cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" exitCode=0 Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.072899 5018 generic.go:334] "Generic (PLEG): container finished" podID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerID="89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" exitCode=143 Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.073004 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.074142 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerDied","Data":"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b"} Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.074193 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerDied","Data":"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857"} Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.074208 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ffc19213-da31-4ee6-b0db-5fa87acfdc45","Type":"ContainerDied","Data":"c674d15dfcf8da8bf7586e941a69d227db8465b246d3504d6fcfb5e811e6b5ff"} Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.074229 5018 scope.go:117] "RemoveContainer" containerID="cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.077438 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-log" containerID="cri-o://935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" gracePeriod=30 Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.077729 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-httpd" containerID="cri-o://e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" gracePeriod=30 Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.077727 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerStarted","Data":"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9"} Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.102058 5018 scope.go:117] "RemoveContainer" containerID="89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.120127 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.120107706 podStartE2EDuration="4.120107706s" podCreationTimestamp="2025-10-14 08:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:49:50.112685453 +0000 UTC m=+7206.696732080" watchObservedRunningTime="2025-10-14 08:49:50.120107706 +0000 UTC m=+7206.704154333" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.127291 5018 scope.go:117] "RemoveContainer" containerID="cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" Oct 14 08:49:50 crc kubenswrapper[5018]: E1014 08:49:50.127892 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b\": container with ID starting with cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b not found: ID does not exist" containerID="cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.127947 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b"} err="failed to get container status \"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b\": rpc error: code = NotFound desc = could not find container \"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b\": container with ID starting with cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b not found: ID does not exist" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.127979 5018 scope.go:117] "RemoveContainer" containerID="89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" Oct 14 08:49:50 crc kubenswrapper[5018]: E1014 08:49:50.128318 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857\": container with ID starting with 89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857 not found: ID does not exist" containerID="89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.128354 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857"} err="failed to get container status \"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857\": rpc error: code = NotFound desc = could not find container \"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857\": container with ID starting with 89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857 not found: ID does not exist" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.128375 5018 scope.go:117] "RemoveContainer" containerID="cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.128749 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b"} err="failed to get container status \"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b\": rpc error: code = NotFound desc = could not find container \"cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b\": container with ID starting with cfc80dbe974f34156a57cfcad29c1408e351edcb770ac6ae48bd43e7e315a16b not found: ID does not exist" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.128806 5018 scope.go:117] "RemoveContainer" containerID="89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.131435 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857"} err="failed to get container status \"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857\": rpc error: code = NotFound desc = could not find container \"89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857\": container with ID starting with 89b5586b0836cea3e11d92387dd3bf97dc3881f4b71c636d9f3cdd1216ba7857 not found: ID does not exist" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.140765 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.167267 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.197237 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:50 crc kubenswrapper[5018]: E1014 08:49:50.197735 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-log" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.197757 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-log" Oct 14 08:49:50 crc kubenswrapper[5018]: E1014 08:49:50.197777 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-httpd" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.197785 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-httpd" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.198036 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-log" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.198067 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" containerName="glance-httpd" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.203371 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.203527 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.205455 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.205640 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.265711 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.265759 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.265874 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.265937 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.266047 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.266072 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.266138 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45754\" (UniqueName: \"kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367232 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367292 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367373 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367399 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367558 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367585 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.367638 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45754\" (UniqueName: \"kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.368428 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.368532 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.373592 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.374154 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.374219 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.374421 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.398302 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45754\" (UniqueName: \"kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754\") pod \"glance-default-external-api-0\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.535823 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.604906 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:49:50 crc kubenswrapper[5018]: E1014 08:49:50.605252 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.619004 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc19213-da31-4ee6-b0db-5fa87acfdc45" path="/var/lib/kubelet/pods/ffc19213-da31-4ee6-b0db-5fa87acfdc45/volumes" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.727738 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.774262 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.774332 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.774371 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.774443 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs7fb\" (UniqueName: \"kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.776027 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs" (OuterVolumeSpecName: "logs") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.776078 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.780172 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb" (OuterVolumeSpecName: "kube-api-access-fs7fb") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "kube-api-access-fs7fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.829549 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data" (OuterVolumeSpecName: "config-data") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.876473 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.876523 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts\") pod \"9a1201d6-932a-4361-a582-22a3a30b699e\" (UID: \"9a1201d6-932a-4361-a582-22a3a30b699e\") " Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.877035 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs7fb\" (UniqueName: \"kubernetes.io/projected/9a1201d6-932a-4361-a582-22a3a30b699e-kube-api-access-fs7fb\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.877059 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.877069 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.877082 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a1201d6-932a-4361-a582-22a3a30b699e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.881566 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts" (OuterVolumeSpecName: "scripts") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.923442 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a1201d6-932a-4361-a582-22a3a30b699e" (UID: "9a1201d6-932a-4361-a582-22a3a30b699e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.978389 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:50 crc kubenswrapper[5018]: I1014 08:49:50.978421 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a1201d6-932a-4361-a582-22a3a30b699e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.087950 5018 generic.go:334] "Generic (PLEG): container finished" podID="9a1201d6-932a-4361-a582-22a3a30b699e" containerID="e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" exitCode=0 Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.087984 5018 generic.go:334] "Generic (PLEG): container finished" podID="9a1201d6-932a-4361-a582-22a3a30b699e" containerID="935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" exitCode=143 Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.088008 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerDied","Data":"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9"} Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.088063 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerDied","Data":"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca"} Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.088081 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a1201d6-932a-4361-a582-22a3a30b699e","Type":"ContainerDied","Data":"e47251690103934223413685b99f1eae9ad15de8667464033a979884bca956a4"} Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.088080 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.088115 5018 scope.go:117] "RemoveContainer" containerID="e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.114356 5018 scope.go:117] "RemoveContainer" containerID="935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.134141 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.142660 5018 scope.go:117] "RemoveContainer" containerID="e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" Oct 14 08:49:51 crc kubenswrapper[5018]: E1014 08:49:51.143240 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9\": container with ID starting with e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9 not found: ID does not exist" containerID="e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.143280 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9"} err="failed to get container status \"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9\": rpc error: code = NotFound desc = could not find container \"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9\": container with ID starting with e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9 not found: ID does not exist" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.143305 5018 scope.go:117] "RemoveContainer" containerID="935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" Oct 14 08:49:51 crc kubenswrapper[5018]: E1014 08:49:51.143943 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca\": container with ID starting with 935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca not found: ID does not exist" containerID="935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.143969 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca"} err="failed to get container status \"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca\": rpc error: code = NotFound desc = could not find container \"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca\": container with ID starting with 935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca not found: ID does not exist" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.143983 5018 scope.go:117] "RemoveContainer" containerID="e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.145118 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9"} err="failed to get container status \"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9\": rpc error: code = NotFound desc = could not find container \"e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9\": container with ID starting with e25d64d2e1955406aac3e23d0892e75f3dd4912fce46d40691167b50ed5ac4b9 not found: ID does not exist" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.145139 5018 scope.go:117] "RemoveContainer" containerID="935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.145449 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.153364 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca"} err="failed to get container status \"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca\": rpc error: code = NotFound desc = could not find container \"935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca\": container with ID starting with 935fd727f2753cb47df14cf9ea2de340c60804dbfcdbb0570702e6be37a85eca not found: ID does not exist" Oct 14 08:49:51 crc kubenswrapper[5018]: W1014 08:49:51.157122 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50595d02_681b_4587_83ce_66e5d5c69951.slice/crio-79c024816c81de2ef7a8fd22e80c3a34a1f041cd3f1c0e336e6e1b81c5488052 WatchSource:0}: Error finding container 79c024816c81de2ef7a8fd22e80c3a34a1f041cd3f1c0e336e6e1b81c5488052: Status 404 returned error can't find the container with id 79c024816c81de2ef7a8fd22e80c3a34a1f041cd3f1c0e336e6e1b81c5488052 Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.159750 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.170484 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:51 crc kubenswrapper[5018]: E1014 08:49:51.170948 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-httpd" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.170969 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-httpd" Oct 14 08:49:51 crc kubenswrapper[5018]: E1014 08:49:51.170987 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-log" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.170995 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-log" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.171226 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-httpd" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.171251 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" containerName="glance-log" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.172390 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.176634 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.177577 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.177862 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.181835 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.181892 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.181930 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.181990 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqj9w\" (UniqueName: \"kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.182089 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.182131 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.182168 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283808 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283856 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283886 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283953 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283978 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.283997 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.284030 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqj9w\" (UniqueName: \"kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.284780 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.287572 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.288237 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.288337 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.289172 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.289997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.306568 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqj9w\" (UniqueName: \"kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w\") pod \"glance-default-internal-api-0\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:49:51 crc kubenswrapper[5018]: I1014 08:49:51.499142 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:49:52 crc kubenswrapper[5018]: I1014 08:49:52.064868 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:49:52 crc kubenswrapper[5018]: I1014 08:49:52.097861 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerStarted","Data":"ca3f3cbc9456e3b957114c2cff5dd46d4b2ff7159e09a6329939bab86aa78bad"} Oct 14 08:49:52 crc kubenswrapper[5018]: I1014 08:49:52.100882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerStarted","Data":"1a6faf8f27efe70c6e8bc2976dade19360f33221cf52aa54557011791f8c78a8"} Oct 14 08:49:52 crc kubenswrapper[5018]: I1014 08:49:52.100930 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerStarted","Data":"79c024816c81de2ef7a8fd22e80c3a34a1f041cd3f1c0e336e6e1b81c5488052"} Oct 14 08:49:52 crc kubenswrapper[5018]: I1014 08:49:52.625210 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a1201d6-932a-4361-a582-22a3a30b699e" path="/var/lib/kubelet/pods/9a1201d6-932a-4361-a582-22a3a30b699e/volumes" Oct 14 08:49:53 crc kubenswrapper[5018]: I1014 08:49:53.117326 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerStarted","Data":"41cc3322c56458de6f5bcf72edcc9e797ada50be777ccec9a283724dbd865ee8"} Oct 14 08:49:53 crc kubenswrapper[5018]: I1014 08:49:53.122032 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerStarted","Data":"c033e84e1d5c677e33aa151740e1f8f7eda2848c801a1b235d43faa5289af0f2"} Oct 14 08:49:53 crc kubenswrapper[5018]: I1014 08:49:53.158885 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.158856078 podStartE2EDuration="3.158856078s" podCreationTimestamp="2025-10-14 08:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:49:53.153747951 +0000 UTC m=+7209.737794608" watchObservedRunningTime="2025-10-14 08:49:53.158856078 +0000 UTC m=+7209.742902745" Oct 14 08:49:54 crc kubenswrapper[5018]: I1014 08:49:54.139875 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerStarted","Data":"295a83f13373ddcea2ff47866b73fdb6d1f09e67500d56c6ab48553e5df24d71"} Oct 14 08:49:54 crc kubenswrapper[5018]: I1014 08:49:54.182185 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.182156734 podStartE2EDuration="3.182156734s" podCreationTimestamp="2025-10-14 08:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:49:54.167305887 +0000 UTC m=+7210.751352554" watchObservedRunningTime="2025-10-14 08:49:54.182156734 +0000 UTC m=+7210.766203371" Oct 14 08:49:56 crc kubenswrapper[5018]: I1014 08:49:56.762646 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:49:56 crc kubenswrapper[5018]: I1014 08:49:56.858113 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:49:56 crc kubenswrapper[5018]: I1014 08:49:56.858565 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="dnsmasq-dns" containerID="cri-o://efd60405ff584a3651531e9f90a1f89f5dcc5641baaff03e8db0c27e7d450878" gracePeriod=10 Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.182076 5018 generic.go:334] "Generic (PLEG): container finished" podID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerID="efd60405ff584a3651531e9f90a1f89f5dcc5641baaff03e8db0c27e7d450878" exitCode=0 Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.182137 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" event={"ID":"56caaa8c-1769-4715-a5e4-ec385d2d6f01","Type":"ContainerDied","Data":"efd60405ff584a3651531e9f90a1f89f5dcc5641baaff03e8db0c27e7d450878"} Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.339007 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.529095 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8mv6\" (UniqueName: \"kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6\") pod \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.529644 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config\") pod \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.529723 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb\") pod \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.529971 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb\") pod \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.530077 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc\") pod \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\" (UID: \"56caaa8c-1769-4715-a5e4-ec385d2d6f01\") " Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.536707 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6" (OuterVolumeSpecName: "kube-api-access-x8mv6") pod "56caaa8c-1769-4715-a5e4-ec385d2d6f01" (UID: "56caaa8c-1769-4715-a5e4-ec385d2d6f01"). InnerVolumeSpecName "kube-api-access-x8mv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.576876 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config" (OuterVolumeSpecName: "config") pod "56caaa8c-1769-4715-a5e4-ec385d2d6f01" (UID: "56caaa8c-1769-4715-a5e4-ec385d2d6f01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.596085 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "56caaa8c-1769-4715-a5e4-ec385d2d6f01" (UID: "56caaa8c-1769-4715-a5e4-ec385d2d6f01"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.599527 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "56caaa8c-1769-4715-a5e4-ec385d2d6f01" (UID: "56caaa8c-1769-4715-a5e4-ec385d2d6f01"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.612181 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "56caaa8c-1769-4715-a5e4-ec385d2d6f01" (UID: "56caaa8c-1769-4715-a5e4-ec385d2d6f01"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.632322 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.632354 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.632364 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8mv6\" (UniqueName: \"kubernetes.io/projected/56caaa8c-1769-4715-a5e4-ec385d2d6f01-kube-api-access-x8mv6\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.632376 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:57 crc kubenswrapper[5018]: I1014 08:49:57.632387 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/56caaa8c-1769-4715-a5e4-ec385d2d6f01-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.200164 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" event={"ID":"56caaa8c-1769-4715-a5e4-ec385d2d6f01","Type":"ContainerDied","Data":"9bb629657c7eca9b449456dd9337c4b6e6738dacb5a51e40c4ef51b480fd67ac"} Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.200242 5018 scope.go:117] "RemoveContainer" containerID="efd60405ff584a3651531e9f90a1f89f5dcc5641baaff03e8db0c27e7d450878" Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.200964 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9c4b98475-w5dsc" Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.249758 5018 scope.go:117] "RemoveContainer" containerID="cfeed0d822d12409d089b725976adacba4ca89b9c79a4c2f02243af9ead8618b" Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.272742 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.283056 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9c4b98475-w5dsc"] Oct 14 08:49:58 crc kubenswrapper[5018]: I1014 08:49:58.618447 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" path="/var/lib/kubelet/pods/56caaa8c-1769-4715-a5e4-ec385d2d6f01/volumes" Oct 14 08:50:00 crc kubenswrapper[5018]: I1014 08:50:00.536353 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:50:00 crc kubenswrapper[5018]: I1014 08:50:00.536742 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:50:00 crc kubenswrapper[5018]: I1014 08:50:00.568503 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:50:00 crc kubenswrapper[5018]: I1014 08:50:00.624711 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.241860 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.242217 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.500012 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.500331 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.578651 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.578754 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:01 crc kubenswrapper[5018]: I1014 08:50:01.608647 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:50:01 crc kubenswrapper[5018]: E1014 08:50:01.623518 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:50:02 crc kubenswrapper[5018]: I1014 08:50:02.258095 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:02 crc kubenswrapper[5018]: I1014 08:50:02.258166 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:03 crc kubenswrapper[5018]: I1014 08:50:03.454222 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:50:03 crc kubenswrapper[5018]: I1014 08:50:03.454970 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:50:03 crc kubenswrapper[5018]: I1014 08:50:03.456474 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:50:04 crc kubenswrapper[5018]: I1014 08:50:04.188363 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:04 crc kubenswrapper[5018]: I1014 08:50:04.192139 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.198922 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-527dq"] Oct 14 08:50:12 crc kubenswrapper[5018]: E1014 08:50:12.199929 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="dnsmasq-dns" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.199947 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="dnsmasq-dns" Oct 14 08:50:12 crc kubenswrapper[5018]: E1014 08:50:12.199999 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="init" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.200007 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="init" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.200243 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="56caaa8c-1769-4715-a5e4-ec385d2d6f01" containerName="dnsmasq-dns" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.201059 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-527dq" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.207660 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-527dq"] Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.366577 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk5cg\" (UniqueName: \"kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg\") pod \"placement-db-create-527dq\" (UID: \"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf\") " pod="openstack/placement-db-create-527dq" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.468840 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk5cg\" (UniqueName: \"kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg\") pod \"placement-db-create-527dq\" (UID: \"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf\") " pod="openstack/placement-db-create-527dq" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.486469 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk5cg\" (UniqueName: \"kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg\") pod \"placement-db-create-527dq\" (UID: \"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf\") " pod="openstack/placement-db-create-527dq" Oct 14 08:50:12 crc kubenswrapper[5018]: I1014 08:50:12.528936 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-527dq" Oct 14 08:50:13 crc kubenswrapper[5018]: W1014 08:50:13.052186 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11428a7e_0d78_45ed_b1c2_1cbc90fe51cf.slice/crio-e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c WatchSource:0}: Error finding container e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c: Status 404 returned error can't find the container with id e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c Oct 14 08:50:13 crc kubenswrapper[5018]: I1014 08:50:13.052442 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-527dq"] Oct 14 08:50:13 crc kubenswrapper[5018]: I1014 08:50:13.366443 5018 generic.go:334] "Generic (PLEG): container finished" podID="11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" containerID="5391d3a26510d78f54511f3524ac18659f17e980c997ac94cdfe99f6c3216466" exitCode=0 Oct 14 08:50:13 crc kubenswrapper[5018]: I1014 08:50:13.366497 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-527dq" event={"ID":"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf","Type":"ContainerDied","Data":"5391d3a26510d78f54511f3524ac18659f17e980c997ac94cdfe99f6c3216466"} Oct 14 08:50:13 crc kubenswrapper[5018]: I1014 08:50:13.366549 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-527dq" event={"ID":"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf","Type":"ContainerStarted","Data":"e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c"} Oct 14 08:50:14 crc kubenswrapper[5018]: I1014 08:50:14.814962 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-527dq" Oct 14 08:50:14 crc kubenswrapper[5018]: I1014 08:50:14.936387 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk5cg\" (UniqueName: \"kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg\") pod \"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf\" (UID: \"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf\") " Oct 14 08:50:14 crc kubenswrapper[5018]: I1014 08:50:14.944100 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg" (OuterVolumeSpecName: "kube-api-access-jk5cg") pod "11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" (UID: "11428a7e-0d78-45ed-b1c2-1cbc90fe51cf"). InnerVolumeSpecName "kube-api-access-jk5cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:15 crc kubenswrapper[5018]: I1014 08:50:15.039086 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk5cg\" (UniqueName: \"kubernetes.io/projected/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf-kube-api-access-jk5cg\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:15 crc kubenswrapper[5018]: I1014 08:50:15.401086 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-527dq" event={"ID":"11428a7e-0d78-45ed-b1c2-1cbc90fe51cf","Type":"ContainerDied","Data":"e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c"} Oct 14 08:50:15 crc kubenswrapper[5018]: I1014 08:50:15.401161 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0677b272e5a17865160405f6332b8f21cd28d8eea28cce7864c38e031e8219c" Oct 14 08:50:15 crc kubenswrapper[5018]: I1014 08:50:15.401587 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-527dq" Oct 14 08:50:16 crc kubenswrapper[5018]: I1014 08:50:16.606949 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:50:16 crc kubenswrapper[5018]: E1014 08:50:16.607820 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.297195 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-85f4-account-create-chbwd"] Oct 14 08:50:22 crc kubenswrapper[5018]: E1014 08:50:22.298755 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" containerName="mariadb-database-create" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.298788 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" containerName="mariadb-database-create" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.299230 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" containerName="mariadb-database-create" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.300520 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.303536 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.307404 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85f4-account-create-chbwd"] Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.401286 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j276v\" (UniqueName: \"kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v\") pod \"placement-85f4-account-create-chbwd\" (UID: \"485a9384-0ea8-4df7-922f-caffb6440309\") " pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.504159 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j276v\" (UniqueName: \"kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v\") pod \"placement-85f4-account-create-chbwd\" (UID: \"485a9384-0ea8-4df7-922f-caffb6440309\") " pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.538062 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j276v\" (UniqueName: \"kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v\") pod \"placement-85f4-account-create-chbwd\" (UID: \"485a9384-0ea8-4df7-922f-caffb6440309\") " pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.631564 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:22 crc kubenswrapper[5018]: I1014 08:50:22.992335 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85f4-account-create-chbwd"] Oct 14 08:50:23 crc kubenswrapper[5018]: I1014 08:50:23.491360 5018 generic.go:334] "Generic (PLEG): container finished" podID="485a9384-0ea8-4df7-922f-caffb6440309" containerID="3511b3411f0b272e1d3c9f02eb319b73fb65032a2dda5b0a0d85238a7ca5bdcd" exitCode=0 Oct 14 08:50:23 crc kubenswrapper[5018]: I1014 08:50:23.491415 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85f4-account-create-chbwd" event={"ID":"485a9384-0ea8-4df7-922f-caffb6440309","Type":"ContainerDied","Data":"3511b3411f0b272e1d3c9f02eb319b73fb65032a2dda5b0a0d85238a7ca5bdcd"} Oct 14 08:50:23 crc kubenswrapper[5018]: I1014 08:50:23.491690 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85f4-account-create-chbwd" event={"ID":"485a9384-0ea8-4df7-922f-caffb6440309","Type":"ContainerStarted","Data":"47c38f952b040c7ba5826f5e72284942b9b2953c7a9b21a0d9d3d24250c2d69d"} Oct 14 08:50:24 crc kubenswrapper[5018]: I1014 08:50:24.934147 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.068158 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j276v\" (UniqueName: \"kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v\") pod \"485a9384-0ea8-4df7-922f-caffb6440309\" (UID: \"485a9384-0ea8-4df7-922f-caffb6440309\") " Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.076831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v" (OuterVolumeSpecName: "kube-api-access-j276v") pod "485a9384-0ea8-4df7-922f-caffb6440309" (UID: "485a9384-0ea8-4df7-922f-caffb6440309"). InnerVolumeSpecName "kube-api-access-j276v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.171484 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j276v\" (UniqueName: \"kubernetes.io/projected/485a9384-0ea8-4df7-922f-caffb6440309-kube-api-access-j276v\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.521140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85f4-account-create-chbwd" event={"ID":"485a9384-0ea8-4df7-922f-caffb6440309","Type":"ContainerDied","Data":"47c38f952b040c7ba5826f5e72284942b9b2953c7a9b21a0d9d3d24250c2d69d"} Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.521201 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47c38f952b040c7ba5826f5e72284942b9b2953c7a9b21a0d9d3d24250c2d69d" Oct 14 08:50:25 crc kubenswrapper[5018]: I1014 08:50:25.521204 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85f4-account-create-chbwd" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.568042 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-lk42t"] Oct 14 08:50:27 crc kubenswrapper[5018]: E1014 08:50:27.568794 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="485a9384-0ea8-4df7-922f-caffb6440309" containerName="mariadb-account-create" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.568812 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="485a9384-0ea8-4df7-922f-caffb6440309" containerName="mariadb-account-create" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.569042 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="485a9384-0ea8-4df7-922f-caffb6440309" containerName="mariadb-account-create" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.569877 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.573949 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.574167 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.574277 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rm7d5" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.587428 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lk42t"] Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.605760 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.610774 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.628315 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731237 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731309 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731493 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731589 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731717 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fp7\" (UniqueName: \"kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731754 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731839 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.731973 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.732077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nc2d\" (UniqueName: \"kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.732151 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.833952 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834035 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nc2d\" (UniqueName: \"kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834069 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834131 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834152 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834194 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834217 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834263 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fp7\" (UniqueName: \"kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834287 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834350 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.834766 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.835539 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.835693 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.835760 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.835840 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.838923 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.840957 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.841222 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.861197 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fp7\" (UniqueName: \"kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7\") pod \"dnsmasq-dns-9759f875c-7kzp6\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.862101 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nc2d\" (UniqueName: \"kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d\") pod \"placement-db-sync-lk42t\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.899946 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:27 crc kubenswrapper[5018]: I1014 08:50:27.945369 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:28 crc kubenswrapper[5018]: I1014 08:50:28.393873 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lk42t"] Oct 14 08:50:28 crc kubenswrapper[5018]: I1014 08:50:28.494051 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:50:28 crc kubenswrapper[5018]: W1014 08:50:28.518986 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298fed88_8ba4_44fa_9327_ec27e95865b4.slice/crio-f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d WatchSource:0}: Error finding container f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d: Status 404 returned error can't find the container with id f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d Oct 14 08:50:28 crc kubenswrapper[5018]: I1014 08:50:28.568147 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lk42t" event={"ID":"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd","Type":"ContainerStarted","Data":"d358c3a18993f9ae8091ec6f89840a73f5e01f1261679c7df26a36443973cd49"} Oct 14 08:50:28 crc kubenswrapper[5018]: I1014 08:50:28.569342 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" event={"ID":"298fed88-8ba4-44fa-9327-ec27e95865b4","Type":"ContainerStarted","Data":"f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d"} Oct 14 08:50:28 crc kubenswrapper[5018]: I1014 08:50:28.605026 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:50:28 crc kubenswrapper[5018]: E1014 08:50:28.605338 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:50:29 crc kubenswrapper[5018]: I1014 08:50:29.585061 5018 generic.go:334] "Generic (PLEG): container finished" podID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerID="d646487382f3976a5b428292b3f200405437c1bf82a45abec7176978973ca91f" exitCode=0 Oct 14 08:50:29 crc kubenswrapper[5018]: I1014 08:50:29.585143 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" event={"ID":"298fed88-8ba4-44fa-9327-ec27e95865b4","Type":"ContainerDied","Data":"d646487382f3976a5b428292b3f200405437c1bf82a45abec7176978973ca91f"} Oct 14 08:50:30 crc kubenswrapper[5018]: I1014 08:50:30.597823 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" event={"ID":"298fed88-8ba4-44fa-9327-ec27e95865b4","Type":"ContainerStarted","Data":"e7ce72802b638ad8af5e63ac212ad5735398a50811daa5c990e45f24450302f6"} Oct 14 08:50:30 crc kubenswrapper[5018]: I1014 08:50:30.598411 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:32 crc kubenswrapper[5018]: I1014 08:50:32.632899 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lk42t" event={"ID":"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd","Type":"ContainerStarted","Data":"0474eb461f962bcb1871b0fbba81d351ab2eccb58550fc06be029fe0f3ec1316"} Oct 14 08:50:32 crc kubenswrapper[5018]: I1014 08:50:32.651495 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" podStartSLOduration=5.6514686130000005 podStartE2EDuration="5.651468613s" podCreationTimestamp="2025-10-14 08:50:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:50:30.618840974 +0000 UTC m=+7247.202887641" watchObservedRunningTime="2025-10-14 08:50:32.651468613 +0000 UTC m=+7249.235515280" Oct 14 08:50:32 crc kubenswrapper[5018]: I1014 08:50:32.654802 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-lk42t" podStartSLOduration=2.295814284 podStartE2EDuration="5.654780698s" podCreationTimestamp="2025-10-14 08:50:27 +0000 UTC" firstStartedPulling="2025-10-14 08:50:28.396918529 +0000 UTC m=+7244.980965176" lastFinishedPulling="2025-10-14 08:50:31.755884963 +0000 UTC m=+7248.339931590" observedRunningTime="2025-10-14 08:50:32.645412469 +0000 UTC m=+7249.229459146" watchObservedRunningTime="2025-10-14 08:50:32.654780698 +0000 UTC m=+7249.238827365" Oct 14 08:50:33 crc kubenswrapper[5018]: I1014 08:50:33.636547 5018 generic.go:334] "Generic (PLEG): container finished" podID="871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" containerID="0474eb461f962bcb1871b0fbba81d351ab2eccb58550fc06be029fe0f3ec1316" exitCode=0 Oct 14 08:50:33 crc kubenswrapper[5018]: I1014 08:50:33.636647 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lk42t" event={"ID":"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd","Type":"ContainerDied","Data":"0474eb461f962bcb1871b0fbba81d351ab2eccb58550fc06be029fe0f3ec1316"} Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.040925 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197000 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle\") pod \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197089 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nc2d\" (UniqueName: \"kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d\") pod \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197127 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs\") pod \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197202 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data\") pod \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197285 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts\") pod \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\" (UID: \"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd\") " Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.197754 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs" (OuterVolumeSpecName: "logs") pod "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" (UID: "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.208965 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d" (OuterVolumeSpecName: "kube-api-access-9nc2d") pod "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" (UID: "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd"). InnerVolumeSpecName "kube-api-access-9nc2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.216763 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts" (OuterVolumeSpecName: "scripts") pod "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" (UID: "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.239690 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" (UID: "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.246785 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data" (OuterVolumeSpecName: "config-data") pod "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" (UID: "871eab4b-2600-4031-a4bd-a8d6e5e1e0bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.299218 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nc2d\" (UniqueName: \"kubernetes.io/projected/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-kube-api-access-9nc2d\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.299248 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.299260 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.299270 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.299278 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.671445 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lk42t" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.671365 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lk42t" event={"ID":"871eab4b-2600-4031-a4bd-a8d6e5e1e0bd","Type":"ContainerDied","Data":"d358c3a18993f9ae8091ec6f89840a73f5e01f1261679c7df26a36443973cd49"} Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.672020 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d358c3a18993f9ae8091ec6f89840a73f5e01f1261679c7df26a36443973cd49" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.875991 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 08:50:35 crc kubenswrapper[5018]: E1014 08:50:35.876485 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" containerName="placement-db-sync" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.876507 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" containerName="placement-db-sync" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.876758 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" containerName="placement-db-sync" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.877996 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.881519 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rm7d5" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.881872 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.884088 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.884333 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.884563 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:50:35 crc kubenswrapper[5018]: I1014 08:50:35.889210 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.018926 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gjhm\" (UniqueName: \"kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019241 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019300 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019328 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019607 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019735 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.019847 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121068 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121126 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121183 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121223 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gjhm\" (UniqueName: \"kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121249 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121288 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121304 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.121980 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.126841 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.127488 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.127839 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.128902 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.132206 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.157248 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gjhm\" (UniqueName: \"kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm\") pod \"placement-5b8df9c49b-fpwlm\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.249552 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:36 crc kubenswrapper[5018]: W1014 08:50:36.764374 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a17b000_a726_4072_8b96_04f317ff7fb0.slice/crio-1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f WatchSource:0}: Error finding container 1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f: Status 404 returned error can't find the container with id 1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f Oct 14 08:50:36 crc kubenswrapper[5018]: I1014 08:50:36.770781 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.695040 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerStarted","Data":"40fb8b68d8bb04c1d344475d4f00e45548e0b4c28983fcb23f3b012d39221907"} Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.695478 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerStarted","Data":"9b40efcd93a346933415044024e90ede001495fcfd8fb0320885bab4c64fda7d"} Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.695501 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerStarted","Data":"1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f"} Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.695531 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.733499 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5b8df9c49b-fpwlm" podStartSLOduration=2.733466409 podStartE2EDuration="2.733466409s" podCreationTimestamp="2025-10-14 08:50:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:50:37.717442369 +0000 UTC m=+7254.301489006" watchObservedRunningTime="2025-10-14 08:50:37.733466409 +0000 UTC m=+7254.317513076" Oct 14 08:50:37 crc kubenswrapper[5018]: I1014 08:50:37.948005 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.053586 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.054006 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="dnsmasq-dns" containerID="cri-o://4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628" gracePeriod=10 Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.617916 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.726782 5018 generic.go:334] "Generic (PLEG): container finished" podID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerID="4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628" exitCode=0 Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.728646 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.729165 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" event={"ID":"94719570-6817-4a41-8d9a-1ce46a4591e8","Type":"ContainerDied","Data":"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628"} Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.729326 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.729542 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c664b6d67-8dmsz" event={"ID":"94719570-6817-4a41-8d9a-1ce46a4591e8","Type":"ContainerDied","Data":"826915519fecd109ddbdc4871a9a7ad2785aa1745111d6c6e884bb7be0fb7d79"} Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.729487 5018 scope.go:117] "RemoveContainer" containerID="4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.751097 5018 scope.go:117] "RemoveContainer" containerID="5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.774888 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config\") pod \"94719570-6817-4a41-8d9a-1ce46a4591e8\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.774997 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb\") pod \"94719570-6817-4a41-8d9a-1ce46a4591e8\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.775065 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb\") pod \"94719570-6817-4a41-8d9a-1ce46a4591e8\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.775111 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgktg\" (UniqueName: \"kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg\") pod \"94719570-6817-4a41-8d9a-1ce46a4591e8\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.775138 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc\") pod \"94719570-6817-4a41-8d9a-1ce46a4591e8\" (UID: \"94719570-6817-4a41-8d9a-1ce46a4591e8\") " Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.777128 5018 scope.go:117] "RemoveContainer" containerID="4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628" Oct 14 08:50:38 crc kubenswrapper[5018]: E1014 08:50:38.778252 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628\": container with ID starting with 4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628 not found: ID does not exist" containerID="4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.778296 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628"} err="failed to get container status \"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628\": rpc error: code = NotFound desc = could not find container \"4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628\": container with ID starting with 4f69776c983501a2cd71cb88eaeb01292e39c42f938b1125abc34df3ed5d1628 not found: ID does not exist" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.778321 5018 scope.go:117] "RemoveContainer" containerID="5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a" Oct 14 08:50:38 crc kubenswrapper[5018]: E1014 08:50:38.779049 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a\": container with ID starting with 5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a not found: ID does not exist" containerID="5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.779128 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a"} err="failed to get container status \"5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a\": rpc error: code = NotFound desc = could not find container \"5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a\": container with ID starting with 5d8ad95ccf7c6b40338b112d1cc380ca3ce103ef1c033d7da63a626ed4a1163a not found: ID does not exist" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.781009 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg" (OuterVolumeSpecName: "kube-api-access-hgktg") pod "94719570-6817-4a41-8d9a-1ce46a4591e8" (UID: "94719570-6817-4a41-8d9a-1ce46a4591e8"). InnerVolumeSpecName "kube-api-access-hgktg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.818964 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94719570-6817-4a41-8d9a-1ce46a4591e8" (UID: "94719570-6817-4a41-8d9a-1ce46a4591e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.827920 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94719570-6817-4a41-8d9a-1ce46a4591e8" (UID: "94719570-6817-4a41-8d9a-1ce46a4591e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.830522 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94719570-6817-4a41-8d9a-1ce46a4591e8" (UID: "94719570-6817-4a41-8d9a-1ce46a4591e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.843980 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config" (OuterVolumeSpecName: "config") pod "94719570-6817-4a41-8d9a-1ce46a4591e8" (UID: "94719570-6817-4a41-8d9a-1ce46a4591e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.876901 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.876936 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.876946 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.876956 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgktg\" (UniqueName: \"kubernetes.io/projected/94719570-6817-4a41-8d9a-1ce46a4591e8-kube-api-access-hgktg\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:38 crc kubenswrapper[5018]: I1014 08:50:38.876964 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94719570-6817-4a41-8d9a-1ce46a4591e8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:39 crc kubenswrapper[5018]: I1014 08:50:39.062089 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:50:39 crc kubenswrapper[5018]: I1014 08:50:39.066347 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c664b6d67-8dmsz"] Oct 14 08:50:39 crc kubenswrapper[5018]: I1014 08:50:39.607321 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:50:39 crc kubenswrapper[5018]: E1014 08:50:39.607785 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:50:40 crc kubenswrapper[5018]: I1014 08:50:40.623043 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" path="/var/lib/kubelet/pods/94719570-6817-4a41-8d9a-1ce46a4591e8/volumes" Oct 14 08:50:51 crc kubenswrapper[5018]: I1014 08:50:51.612343 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:50:51 crc kubenswrapper[5018]: E1014 08:50:51.620380 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:03 crc kubenswrapper[5018]: I1014 08:51:03.605018 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:51:03 crc kubenswrapper[5018]: E1014 08:51:03.605920 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:07 crc kubenswrapper[5018]: I1014 08:51:07.240975 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:51:08 crc kubenswrapper[5018]: I1014 08:51:08.253017 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 08:51:15 crc kubenswrapper[5018]: I1014 08:51:15.605886 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:51:15 crc kubenswrapper[5018]: E1014 08:51:15.606636 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.594460 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:19 crc kubenswrapper[5018]: E1014 08:51:19.595380 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="dnsmasq-dns" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.595398 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="dnsmasq-dns" Oct 14 08:51:19 crc kubenswrapper[5018]: E1014 08:51:19.595434 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="init" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.595443 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="init" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.595696 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="94719570-6817-4a41-8d9a-1ce46a4591e8" containerName="dnsmasq-dns" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.597124 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.603584 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.744197 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.744259 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9tt2\" (UniqueName: \"kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.744434 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.845686 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.845734 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9tt2\" (UniqueName: \"kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.845803 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.846287 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.846302 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.868435 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9tt2\" (UniqueName: \"kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2\") pod \"certified-operators-grkkk\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:19 crc kubenswrapper[5018]: I1014 08:51:19.922549 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:20 crc kubenswrapper[5018]: I1014 08:51:20.435319 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:21 crc kubenswrapper[5018]: I1014 08:51:21.182084 5018 generic.go:334] "Generic (PLEG): container finished" podID="39e2876b-8835-4c47-888e-0ec648179a01" containerID="81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f" exitCode=0 Oct 14 08:51:21 crc kubenswrapper[5018]: I1014 08:51:21.182148 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerDied","Data":"81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f"} Oct 14 08:51:21 crc kubenswrapper[5018]: I1014 08:51:21.182469 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerStarted","Data":"4cc1f22d0a3a2ec8ba48c523062ea3d808f33caa8c9af877628879cc3b769828"} Oct 14 08:51:23 crc kubenswrapper[5018]: I1014 08:51:23.203937 5018 generic.go:334] "Generic (PLEG): container finished" podID="39e2876b-8835-4c47-888e-0ec648179a01" containerID="572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc" exitCode=0 Oct 14 08:51:23 crc kubenswrapper[5018]: I1014 08:51:23.204107 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerDied","Data":"572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc"} Oct 14 08:51:24 crc kubenswrapper[5018]: I1014 08:51:24.216913 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerStarted","Data":"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6"} Oct 14 08:51:24 crc kubenswrapper[5018]: I1014 08:51:24.241707 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-grkkk" podStartSLOduration=2.613150728 podStartE2EDuration="5.241686484s" podCreationTimestamp="2025-10-14 08:51:19 +0000 UTC" firstStartedPulling="2025-10-14 08:51:21.184926885 +0000 UTC m=+7297.768973542" lastFinishedPulling="2025-10-14 08:51:23.813462661 +0000 UTC m=+7300.397509298" observedRunningTime="2025-10-14 08:51:24.233718585 +0000 UTC m=+7300.817765222" watchObservedRunningTime="2025-10-14 08:51:24.241686484 +0000 UTC m=+7300.825733131" Oct 14 08:51:29 crc kubenswrapper[5018]: I1014 08:51:29.605966 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:51:29 crc kubenswrapper[5018]: E1014 08:51:29.607493 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:29 crc kubenswrapper[5018]: I1014 08:51:29.922792 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:29 crc kubenswrapper[5018]: I1014 08:51:29.922852 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:30 crc kubenswrapper[5018]: I1014 08:51:30.003096 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:30 crc kubenswrapper[5018]: I1014 08:51:30.340304 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:30 crc kubenswrapper[5018]: I1014 08:51:30.404737 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.288289 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-grkkk" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="registry-server" containerID="cri-o://2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6" gracePeriod=2 Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.740867 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.853030 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9tt2\" (UniqueName: \"kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2\") pod \"39e2876b-8835-4c47-888e-0ec648179a01\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.853246 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content\") pod \"39e2876b-8835-4c47-888e-0ec648179a01\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.853368 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities\") pod \"39e2876b-8835-4c47-888e-0ec648179a01\" (UID: \"39e2876b-8835-4c47-888e-0ec648179a01\") " Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.855976 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities" (OuterVolumeSpecName: "utilities") pod "39e2876b-8835-4c47-888e-0ec648179a01" (UID: "39e2876b-8835-4c47-888e-0ec648179a01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.860016 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2" (OuterVolumeSpecName: "kube-api-access-v9tt2") pod "39e2876b-8835-4c47-888e-0ec648179a01" (UID: "39e2876b-8835-4c47-888e-0ec648179a01"). InnerVolumeSpecName "kube-api-access-v9tt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.913931 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39e2876b-8835-4c47-888e-0ec648179a01" (UID: "39e2876b-8835-4c47-888e-0ec648179a01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.955854 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9tt2\" (UniqueName: \"kubernetes.io/projected/39e2876b-8835-4c47-888e-0ec648179a01-kube-api-access-v9tt2\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.955884 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:32 crc kubenswrapper[5018]: I1014 08:51:32.955894 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39e2876b-8835-4c47-888e-0ec648179a01-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.238695 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-qrvbw"] Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.239149 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="extract-content" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.239169 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="extract-content" Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.239212 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="registry-server" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.239221 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="registry-server" Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.239234 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="extract-utilities" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.239242 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="extract-utilities" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.239456 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e2876b-8835-4c47-888e-0ec648179a01" containerName="registry-server" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.240195 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.251258 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qrvbw"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.330486 5018 generic.go:334] "Generic (PLEG): container finished" podID="39e2876b-8835-4c47-888e-0ec648179a01" containerID="2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6" exitCode=0 Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.330531 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerDied","Data":"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6"} Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.330561 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grkkk" event={"ID":"39e2876b-8835-4c47-888e-0ec648179a01","Type":"ContainerDied","Data":"4cc1f22d0a3a2ec8ba48c523062ea3d808f33caa8c9af877628879cc3b769828"} Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.330580 5018 scope.go:117] "RemoveContainer" containerID="2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.330773 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grkkk" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.334322 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-m8j5s"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.336054 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.352818 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-m8j5s"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.355816 5018 scope.go:117] "RemoveContainer" containerID="572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.394135 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.396956 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn76z\" (UniqueName: \"kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z\") pod \"nova-api-db-create-qrvbw\" (UID: \"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f\") " pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.399360 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-grkkk"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.418077 5018 scope.go:117] "RemoveContainer" containerID="81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.426411 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-p2mrn"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.427859 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.434037 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p2mrn"] Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.452136 5018 scope.go:117] "RemoveContainer" containerID="2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6" Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.453077 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6\": container with ID starting with 2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6 not found: ID does not exist" containerID="2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.453198 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6"} err="failed to get container status \"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6\": rpc error: code = NotFound desc = could not find container \"2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6\": container with ID starting with 2facc1f3d5a857cb06970f635f6826d95cf5e67cd19e45541245d08d06c26cb6 not found: ID does not exist" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.453290 5018 scope.go:117] "RemoveContainer" containerID="572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc" Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.453783 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc\": container with ID starting with 572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc not found: ID does not exist" containerID="572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.453908 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc"} err="failed to get container status \"572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc\": rpc error: code = NotFound desc = could not find container \"572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc\": container with ID starting with 572d9c1dcb595f697311896aacfdde8aba3dbb01ac6e336e7e9cbb3e210a0cfc not found: ID does not exist" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.453945 5018 scope.go:117] "RemoveContainer" containerID="81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f" Oct 14 08:51:33 crc kubenswrapper[5018]: E1014 08:51:33.455682 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f\": container with ID starting with 81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f not found: ID does not exist" containerID="81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.455712 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f"} err="failed to get container status \"81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f\": rpc error: code = NotFound desc = could not find container \"81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f\": container with ID starting with 81fe59608f863143e09948f529399be9141767ba978ba1a857624bdeb542b57f not found: ID does not exist" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.500508 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8sbk\" (UniqueName: \"kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk\") pod \"nova-cell1-db-create-p2mrn\" (UID: \"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8\") " pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.500555 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdblz\" (UniqueName: \"kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz\") pod \"nova-cell0-db-create-m8j5s\" (UID: \"0ab5c795-ddf6-4f81-bd9e-f15389bbba38\") " pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.500596 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn76z\" (UniqueName: \"kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z\") pod \"nova-api-db-create-qrvbw\" (UID: \"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f\") " pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.517223 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn76z\" (UniqueName: \"kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z\") pod \"nova-api-db-create-qrvbw\" (UID: \"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f\") " pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.602270 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdblz\" (UniqueName: \"kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz\") pod \"nova-cell0-db-create-m8j5s\" (UID: \"0ab5c795-ddf6-4f81-bd9e-f15389bbba38\") " pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.602444 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8sbk\" (UniqueName: \"kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk\") pod \"nova-cell1-db-create-p2mrn\" (UID: \"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8\") " pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.619418 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8sbk\" (UniqueName: \"kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk\") pod \"nova-cell1-db-create-p2mrn\" (UID: \"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8\") " pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.620083 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdblz\" (UniqueName: \"kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz\") pod \"nova-cell0-db-create-m8j5s\" (UID: \"0ab5c795-ddf6-4f81-bd9e-f15389bbba38\") " pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.630669 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.660077 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:33 crc kubenswrapper[5018]: I1014 08:51:33.779484 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.126802 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qrvbw"] Oct 14 08:51:34 crc kubenswrapper[5018]: W1014 08:51:34.128939 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6494b4a3_ba5b_4eda_9825_7f5b3acefd5f.slice/crio-d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3 WatchSource:0}: Error finding container d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3: Status 404 returned error can't find the container with id d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3 Oct 14 08:51:34 crc kubenswrapper[5018]: W1014 08:51:34.177867 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ab5c795_ddf6_4f81_bd9e_f15389bbba38.slice/crio-a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996 WatchSource:0}: Error finding container a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996: Status 404 returned error can't find the container with id a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996 Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.178924 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-m8j5s"] Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.238526 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p2mrn"] Oct 14 08:51:34 crc kubenswrapper[5018]: W1014 08:51:34.261315 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b05444e_ffec_436a_bb1f_1c1b31ebf7a8.slice/crio-3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87 WatchSource:0}: Error finding container 3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87: Status 404 returned error can't find the container with id 3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87 Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.345226 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p2mrn" event={"ID":"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8","Type":"ContainerStarted","Data":"3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87"} Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.351500 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-m8j5s" event={"ID":"0ab5c795-ddf6-4f81-bd9e-f15389bbba38","Type":"ContainerStarted","Data":"a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996"} Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.362912 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qrvbw" event={"ID":"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f","Type":"ContainerStarted","Data":"d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3"} Oct 14 08:51:34 crc kubenswrapper[5018]: I1014 08:51:34.616398 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e2876b-8835-4c47-888e-0ec648179a01" path="/var/lib/kubelet/pods/39e2876b-8835-4c47-888e-0ec648179a01/volumes" Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.379451 5018 generic.go:334] "Generic (PLEG): container finished" podID="7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" containerID="bd1e1e82e5ba7088d371b42b80db29cf7e78000b1ebbcd9beb7cafe27869faa3" exitCode=0 Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.379648 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p2mrn" event={"ID":"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8","Type":"ContainerDied","Data":"bd1e1e82e5ba7088d371b42b80db29cf7e78000b1ebbcd9beb7cafe27869faa3"} Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.383853 5018 generic.go:334] "Generic (PLEG): container finished" podID="0ab5c795-ddf6-4f81-bd9e-f15389bbba38" containerID="3b4139a5719d9a5784da94eef162c449f1bc1776b8058820ae822f1d0f8f4ce1" exitCode=0 Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.383935 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-m8j5s" event={"ID":"0ab5c795-ddf6-4f81-bd9e-f15389bbba38","Type":"ContainerDied","Data":"3b4139a5719d9a5784da94eef162c449f1bc1776b8058820ae822f1d0f8f4ce1"} Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.387689 5018 generic.go:334] "Generic (PLEG): container finished" podID="6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" containerID="b524bc97d5ef2425dbc03e111632232af0f5a589133c84eec564c80c53abb4af" exitCode=0 Oct 14 08:51:35 crc kubenswrapper[5018]: I1014 08:51:35.387747 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qrvbw" event={"ID":"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f","Type":"ContainerDied","Data":"b524bc97d5ef2425dbc03e111632232af0f5a589133c84eec564c80c53abb4af"} Oct 14 08:51:36 crc kubenswrapper[5018]: I1014 08:51:36.883163 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:36 crc kubenswrapper[5018]: I1014 08:51:36.888440 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:36 crc kubenswrapper[5018]: I1014 08:51:36.956232 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.073775 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdblz\" (UniqueName: \"kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz\") pod \"0ab5c795-ddf6-4f81-bd9e-f15389bbba38\" (UID: \"0ab5c795-ddf6-4f81-bd9e-f15389bbba38\") " Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.074716 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8sbk\" (UniqueName: \"kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk\") pod \"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8\" (UID: \"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8\") " Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.074838 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn76z\" (UniqueName: \"kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z\") pod \"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f\" (UID: \"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f\") " Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.082968 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z" (OuterVolumeSpecName: "kube-api-access-sn76z") pod "6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" (UID: "6494b4a3-ba5b-4eda-9825-7f5b3acefd5f"). InnerVolumeSpecName "kube-api-access-sn76z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.083014 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz" (OuterVolumeSpecName: "kube-api-access-qdblz") pod "0ab5c795-ddf6-4f81-bd9e-f15389bbba38" (UID: "0ab5c795-ddf6-4f81-bd9e-f15389bbba38"). InnerVolumeSpecName "kube-api-access-qdblz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.085861 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk" (OuterVolumeSpecName: "kube-api-access-m8sbk") pod "7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" (UID: "7b05444e-ffec-436a-bb1f-1c1b31ebf7a8"). InnerVolumeSpecName "kube-api-access-m8sbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.177235 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8sbk\" (UniqueName: \"kubernetes.io/projected/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8-kube-api-access-m8sbk\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.177307 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn76z\" (UniqueName: \"kubernetes.io/projected/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f-kube-api-access-sn76z\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.177320 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdblz\" (UniqueName: \"kubernetes.io/projected/0ab5c795-ddf6-4f81-bd9e-f15389bbba38-kube-api-access-qdblz\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.415125 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p2mrn" event={"ID":"7b05444e-ffec-436a-bb1f-1c1b31ebf7a8","Type":"ContainerDied","Data":"3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87"} Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.415222 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3009b23cf7e9eea6457d6c74fcb51204988e2c36327e0cde2d91da2f496b9f87" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.415396 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p2mrn" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.418198 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-m8j5s" event={"ID":"0ab5c795-ddf6-4f81-bd9e-f15389bbba38","Type":"ContainerDied","Data":"a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996"} Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.418264 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8e05969d914ebd8e5d308b92f4b44704340761670699073ff65947fa5184996" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.418313 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-m8j5s" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.420655 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qrvbw" event={"ID":"6494b4a3-ba5b-4eda-9825-7f5b3acefd5f","Type":"ContainerDied","Data":"d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3"} Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.420710 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6b9708c40e48e6ee6f90c3b50754197a03ba7b560997aea4b78ce2cafbb59f3" Oct 14 08:51:37 crc kubenswrapper[5018]: I1014 08:51:37.420730 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qrvbw" Oct 14 08:51:41 crc kubenswrapper[5018]: I1014 08:51:41.605235 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:51:41 crc kubenswrapper[5018]: E1014 08:51:41.606346 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.387149 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-aea5-account-create-gfrn8"] Oct 14 08:51:43 crc kubenswrapper[5018]: E1014 08:51:43.387989 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388012 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: E1014 08:51:43.388043 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388054 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: E1014 08:51:43.388084 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab5c795-ddf6-4f81-bd9e-f15389bbba38" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388094 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab5c795-ddf6-4f81-bd9e-f15389bbba38" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388315 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab5c795-ddf6-4f81-bd9e-f15389bbba38" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388361 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.388377 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" containerName="mariadb-database-create" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.389171 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.392124 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.403141 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aea5-account-create-gfrn8"] Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.513395 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrxdp\" (UniqueName: \"kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp\") pod \"nova-api-aea5-account-create-gfrn8\" (UID: \"d5c7cdc3-830e-416d-9218-561f257c524a\") " pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.569234 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8298-account-create-jtn5n"] Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.575748 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.580990 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.585471 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8298-account-create-jtn5n"] Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.617245 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrxdp\" (UniqueName: \"kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp\") pod \"nova-api-aea5-account-create-gfrn8\" (UID: \"d5c7cdc3-830e-416d-9218-561f257c524a\") " pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.634712 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrxdp\" (UniqueName: \"kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp\") pod \"nova-api-aea5-account-create-gfrn8\" (UID: \"d5c7cdc3-830e-416d-9218-561f257c524a\") " pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.712082 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.728941 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcrtm\" (UniqueName: \"kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm\") pod \"nova-cell0-8298-account-create-jtn5n\" (UID: \"9461fc57-4388-43b1-84d1-08c4231f8aad\") " pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.736973 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-eeba-account-create-mfhsn"] Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.738176 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.740819 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.747651 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-eeba-account-create-mfhsn"] Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.832228 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcrtm\" (UniqueName: \"kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm\") pod \"nova-cell0-8298-account-create-jtn5n\" (UID: \"9461fc57-4388-43b1-84d1-08c4231f8aad\") " pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.856379 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcrtm\" (UniqueName: \"kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm\") pod \"nova-cell0-8298-account-create-jtn5n\" (UID: \"9461fc57-4388-43b1-84d1-08c4231f8aad\") " pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.896303 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:43 crc kubenswrapper[5018]: I1014 08:51:43.933968 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbnbq\" (UniqueName: \"kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq\") pod \"nova-cell1-eeba-account-create-mfhsn\" (UID: \"02f3934a-36ce-494e-8c56-707748aaf954\") " pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.035446 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbnbq\" (UniqueName: \"kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq\") pod \"nova-cell1-eeba-account-create-mfhsn\" (UID: \"02f3934a-36ce-494e-8c56-707748aaf954\") " pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.055925 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbnbq\" (UniqueName: \"kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq\") pod \"nova-cell1-eeba-account-create-mfhsn\" (UID: \"02f3934a-36ce-494e-8c56-707748aaf954\") " pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.137104 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.168117 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-aea5-account-create-gfrn8"] Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.322133 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8298-account-create-jtn5n"] Oct 14 08:51:44 crc kubenswrapper[5018]: W1014 08:51:44.333180 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9461fc57_4388_43b1_84d1_08c4231f8aad.slice/crio-34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb WatchSource:0}: Error finding container 34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb: Status 404 returned error can't find the container with id 34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.531780 5018 generic.go:334] "Generic (PLEG): container finished" podID="d5c7cdc3-830e-416d-9218-561f257c524a" containerID="694bc6842c5310eb5784443688aa5e6a81785acbd4e872a7bfe75437a6415cb7" exitCode=0 Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.531849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aea5-account-create-gfrn8" event={"ID":"d5c7cdc3-830e-416d-9218-561f257c524a","Type":"ContainerDied","Data":"694bc6842c5310eb5784443688aa5e6a81785acbd4e872a7bfe75437a6415cb7"} Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.532143 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aea5-account-create-gfrn8" event={"ID":"d5c7cdc3-830e-416d-9218-561f257c524a","Type":"ContainerStarted","Data":"c28f5c5a37bbca768d1389bbd37bc93d4d01b5303cb922b4ca79af6cda1c05ef"} Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.534424 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8298-account-create-jtn5n" event={"ID":"9461fc57-4388-43b1-84d1-08c4231f8aad","Type":"ContainerStarted","Data":"bf5e3c508a939f648024a637db2a5b33ea30c5568dc14e1a60d8cc63b19e066c"} Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.534471 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8298-account-create-jtn5n" event={"ID":"9461fc57-4388-43b1-84d1-08c4231f8aad","Type":"ContainerStarted","Data":"34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb"} Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.575846 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8298-account-create-jtn5n" podStartSLOduration=1.5758297460000001 podStartE2EDuration="1.575829746s" podCreationTimestamp="2025-10-14 08:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:51:44.568692941 +0000 UTC m=+7321.152739568" watchObservedRunningTime="2025-10-14 08:51:44.575829746 +0000 UTC m=+7321.159876373" Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.665919 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-eeba-account-create-mfhsn"] Oct 14 08:51:44 crc kubenswrapper[5018]: W1014 08:51:44.670138 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f3934a_36ce_494e_8c56_707748aaf954.slice/crio-dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207 WatchSource:0}: Error finding container dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207: Status 404 returned error can't find the container with id dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207 Oct 14 08:51:44 crc kubenswrapper[5018]: I1014 08:51:44.675811 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.551078 5018 generic.go:334] "Generic (PLEG): container finished" podID="9461fc57-4388-43b1-84d1-08c4231f8aad" containerID="bf5e3c508a939f648024a637db2a5b33ea30c5568dc14e1a60d8cc63b19e066c" exitCode=0 Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.551194 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8298-account-create-jtn5n" event={"ID":"9461fc57-4388-43b1-84d1-08c4231f8aad","Type":"ContainerDied","Data":"bf5e3c508a939f648024a637db2a5b33ea30c5568dc14e1a60d8cc63b19e066c"} Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.554455 5018 generic.go:334] "Generic (PLEG): container finished" podID="02f3934a-36ce-494e-8c56-707748aaf954" containerID="1e0ab488798d47de0be9afd2be9d4121397609cc00e13ec274b3fa8cab4cf8c7" exitCode=0 Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.554564 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-eeba-account-create-mfhsn" event={"ID":"02f3934a-36ce-494e-8c56-707748aaf954","Type":"ContainerDied","Data":"1e0ab488798d47de0be9afd2be9d4121397609cc00e13ec274b3fa8cab4cf8c7"} Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.554737 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-eeba-account-create-mfhsn" event={"ID":"02f3934a-36ce-494e-8c56-707748aaf954","Type":"ContainerStarted","Data":"dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207"} Oct 14 08:51:45 crc kubenswrapper[5018]: I1014 08:51:45.896981 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.076358 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrxdp\" (UniqueName: \"kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp\") pod \"d5c7cdc3-830e-416d-9218-561f257c524a\" (UID: \"d5c7cdc3-830e-416d-9218-561f257c524a\") " Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.088661 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp" (OuterVolumeSpecName: "kube-api-access-wrxdp") pod "d5c7cdc3-830e-416d-9218-561f257c524a" (UID: "d5c7cdc3-830e-416d-9218-561f257c524a"). InnerVolumeSpecName "kube-api-access-wrxdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.180897 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrxdp\" (UniqueName: \"kubernetes.io/projected/d5c7cdc3-830e-416d-9218-561f257c524a-kube-api-access-wrxdp\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.569462 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-aea5-account-create-gfrn8" event={"ID":"d5c7cdc3-830e-416d-9218-561f257c524a","Type":"ContainerDied","Data":"c28f5c5a37bbca768d1389bbd37bc93d4d01b5303cb922b4ca79af6cda1c05ef"} Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.569525 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c28f5c5a37bbca768d1389bbd37bc93d4d01b5303cb922b4ca79af6cda1c05ef" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.569724 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-aea5-account-create-gfrn8" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.971648 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:46 crc kubenswrapper[5018]: I1014 08:51:46.978367 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.096476 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcrtm\" (UniqueName: \"kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm\") pod \"9461fc57-4388-43b1-84d1-08c4231f8aad\" (UID: \"9461fc57-4388-43b1-84d1-08c4231f8aad\") " Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.096789 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbnbq\" (UniqueName: \"kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq\") pod \"02f3934a-36ce-494e-8c56-707748aaf954\" (UID: \"02f3934a-36ce-494e-8c56-707748aaf954\") " Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.100208 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq" (OuterVolumeSpecName: "kube-api-access-bbnbq") pod "02f3934a-36ce-494e-8c56-707748aaf954" (UID: "02f3934a-36ce-494e-8c56-707748aaf954"). InnerVolumeSpecName "kube-api-access-bbnbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.100362 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm" (OuterVolumeSpecName: "kube-api-access-wcrtm") pod "9461fc57-4388-43b1-84d1-08c4231f8aad" (UID: "9461fc57-4388-43b1-84d1-08c4231f8aad"). InnerVolumeSpecName "kube-api-access-wcrtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.199035 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcrtm\" (UniqueName: \"kubernetes.io/projected/9461fc57-4388-43b1-84d1-08c4231f8aad-kube-api-access-wcrtm\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.199086 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbnbq\" (UniqueName: \"kubernetes.io/projected/02f3934a-36ce-494e-8c56-707748aaf954-kube-api-access-bbnbq\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.586601 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-eeba-account-create-mfhsn" event={"ID":"02f3934a-36ce-494e-8c56-707748aaf954","Type":"ContainerDied","Data":"dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207"} Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.586692 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc62c2e7def5e73664a2545b8ea603f4493bc339f77e539078990b8cdf131207" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.586788 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-eeba-account-create-mfhsn" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.588955 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8298-account-create-jtn5n" event={"ID":"9461fc57-4388-43b1-84d1-08c4231f8aad","Type":"ContainerDied","Data":"34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb"} Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.588986 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34cc4821d927f5d2ce69059ea79cd8c5eddae0e74584b2db6db1d00edeb8b8cb" Oct 14 08:51:47 crc kubenswrapper[5018]: I1014 08:51:47.589035 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8298-account-create-jtn5n" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773339 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdg8x"] Oct 14 08:51:48 crc kubenswrapper[5018]: E1014 08:51:48.773731 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c7cdc3-830e-416d-9218-561f257c524a" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773743 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c7cdc3-830e-416d-9218-561f257c524a" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: E1014 08:51:48.773751 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9461fc57-4388-43b1-84d1-08c4231f8aad" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773758 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9461fc57-4388-43b1-84d1-08c4231f8aad" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: E1014 08:51:48.773766 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f3934a-36ce-494e-8c56-707748aaf954" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773772 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f3934a-36ce-494e-8c56-707748aaf954" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773950 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c7cdc3-830e-416d-9218-561f257c524a" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773960 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f3934a-36ce-494e-8c56-707748aaf954" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.773972 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9461fc57-4388-43b1-84d1-08c4231f8aad" containerName="mariadb-account-create" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.774568 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.777400 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-69d67" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.777538 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.783692 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.784029 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdg8x"] Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.935082 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7wjs\" (UniqueName: \"kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.935495 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.935557 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:48 crc kubenswrapper[5018]: I1014 08:51:48.935684 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.037108 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.037246 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7wjs\" (UniqueName: \"kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.037322 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.037354 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.043095 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.043410 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.050569 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.082692 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7wjs\" (UniqueName: \"kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs\") pod \"nova-cell0-conductor-db-sync-fdg8x\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.097724 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.579871 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdg8x"] Oct 14 08:51:49 crc kubenswrapper[5018]: W1014 08:51:49.583818 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44c7e1f8_4d12_4732_8eb9_d99a6bbd9d8c.slice/crio-c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1 WatchSource:0}: Error finding container c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1: Status 404 returned error can't find the container with id c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1 Oct 14 08:51:49 crc kubenswrapper[5018]: I1014 08:51:49.634822 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" event={"ID":"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c","Type":"ContainerStarted","Data":"c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1"} Oct 14 08:51:55 crc kubenswrapper[5018]: I1014 08:51:55.604871 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:51:55 crc kubenswrapper[5018]: E1014 08:51:55.605744 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:51:58 crc kubenswrapper[5018]: I1014 08:51:58.741165 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" event={"ID":"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c","Type":"ContainerStarted","Data":"cdb384054bfc015a9477cd485c9b6d9ebbb34e899544fa170bed94271ce4a0ea"} Oct 14 08:51:58 crc kubenswrapper[5018]: I1014 08:51:58.789366 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" podStartSLOduration=2.633168242 podStartE2EDuration="10.789334456s" podCreationTimestamp="2025-10-14 08:51:48 +0000 UTC" firstStartedPulling="2025-10-14 08:51:49.585680652 +0000 UTC m=+7326.169727289" lastFinishedPulling="2025-10-14 08:51:57.741846846 +0000 UTC m=+7334.325893503" observedRunningTime="2025-10-14 08:51:58.783529119 +0000 UTC m=+7335.367575786" watchObservedRunningTime="2025-10-14 08:51:58.789334456 +0000 UTC m=+7335.373381123" Oct 14 08:52:03 crc kubenswrapper[5018]: I1014 08:52:03.806559 5018 generic.go:334] "Generic (PLEG): container finished" podID="44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" containerID="cdb384054bfc015a9477cd485c9b6d9ebbb34e899544fa170bed94271ce4a0ea" exitCode=0 Oct 14 08:52:03 crc kubenswrapper[5018]: I1014 08:52:03.806718 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" event={"ID":"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c","Type":"ContainerDied","Data":"cdb384054bfc015a9477cd485c9b6d9ebbb34e899544fa170bed94271ce4a0ea"} Oct 14 08:52:04 crc kubenswrapper[5018]: I1014 08:52:04.085746 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-p45q9"] Oct 14 08:52:04 crc kubenswrapper[5018]: I1014 08:52:04.102483 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-p45q9"] Oct 14 08:52:04 crc kubenswrapper[5018]: I1014 08:52:04.625513 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c76f3114-2b8a-4dab-a099-7934a05b33aa" path="/var/lib/kubelet/pods/c76f3114-2b8a-4dab-a099-7934a05b33aa/volumes" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.299411 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.470222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle\") pod \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.470301 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7wjs\" (UniqueName: \"kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs\") pod \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.470476 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data\") pod \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.470499 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts\") pod \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\" (UID: \"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c\") " Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.476578 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts" (OuterVolumeSpecName: "scripts") pod "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" (UID: "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.477500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs" (OuterVolumeSpecName: "kube-api-access-t7wjs") pod "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" (UID: "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c"). InnerVolumeSpecName "kube-api-access-t7wjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.511931 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" (UID: "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.528172 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data" (OuterVolumeSpecName: "config-data") pod "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" (UID: "44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.572853 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.572906 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7wjs\" (UniqueName: \"kubernetes.io/projected/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-kube-api-access-t7wjs\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.572921 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.572932 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.837666 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" event={"ID":"44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c","Type":"ContainerDied","Data":"c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1"} Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.838170 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0c0b0031a951e57160ab7ede8bdde1d8f810dc5549893d17fd05c8bf70e4ec1" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.837743 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdg8x" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.978822 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:52:05 crc kubenswrapper[5018]: E1014 08:52:05.979675 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" containerName="nova-cell0-conductor-db-sync" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.979707 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" containerName="nova-cell0-conductor-db-sync" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.980087 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" containerName="nova-cell0-conductor-db-sync" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.981403 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.985440 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.985752 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-69d67" Oct 14 08:52:05 crc kubenswrapper[5018]: I1014 08:52:05.987106 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.088535 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.088924 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crkkh\" (UniqueName: \"kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.089097 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.191341 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.191489 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crkkh\" (UniqueName: \"kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.191553 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.196933 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.203289 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.212136 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crkkh\" (UniqueName: \"kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh\") pod \"nova-cell0-conductor-0\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.344879 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.605793 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:52:06 crc kubenswrapper[5018]: E1014 08:52:06.606459 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:52:06 crc kubenswrapper[5018]: W1014 08:52:06.882519 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fe75171_60d0_48df_a61a_c0e828f23f66.slice/crio-09267f3128f593fc994473c067db631355597be81a0f9fc87a429b619699ca40 WatchSource:0}: Error finding container 09267f3128f593fc994473c067db631355597be81a0f9fc87a429b619699ca40: Status 404 returned error can't find the container with id 09267f3128f593fc994473c067db631355597be81a0f9fc87a429b619699ca40 Oct 14 08:52:06 crc kubenswrapper[5018]: I1014 08:52:06.883544 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:52:07 crc kubenswrapper[5018]: I1014 08:52:07.861674 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fe75171-60d0-48df-a61a-c0e828f23f66","Type":"ContainerStarted","Data":"43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171"} Oct 14 08:52:07 crc kubenswrapper[5018]: I1014 08:52:07.862348 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fe75171-60d0-48df-a61a-c0e828f23f66","Type":"ContainerStarted","Data":"09267f3128f593fc994473c067db631355597be81a0f9fc87a429b619699ca40"} Oct 14 08:52:07 crc kubenswrapper[5018]: I1014 08:52:07.862408 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:08 crc kubenswrapper[5018]: I1014 08:52:08.972555 5018 scope.go:117] "RemoveContainer" containerID="17158a8fa8c3a15c6b14e71d5b226e53334837e0ea5686edc05334d9ee3734c6" Oct 14 08:52:14 crc kubenswrapper[5018]: I1014 08:52:14.013090 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=9.013068946 podStartE2EDuration="9.013068946s" podCreationTimestamp="2025-10-14 08:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:07.881876432 +0000 UTC m=+7344.465923069" watchObservedRunningTime="2025-10-14 08:52:14.013068946 +0000 UTC m=+7350.597115573" Oct 14 08:52:14 crc kubenswrapper[5018]: I1014 08:52:14.020995 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-919f-account-create-vdsxq"] Oct 14 08:52:14 crc kubenswrapper[5018]: I1014 08:52:14.027840 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-919f-account-create-vdsxq"] Oct 14 08:52:14 crc kubenswrapper[5018]: I1014 08:52:14.619719 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e155fce-f943-4909-a794-3d8c3d2bb036" path="/var/lib/kubelet/pods/5e155fce-f943-4909-a794-3d8c3d2bb036/volumes" Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.398029 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.936498 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ch8tx"] Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.937673 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.946426 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.946576 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 08:52:16 crc kubenswrapper[5018]: I1014 08:52:16.952973 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ch8tx"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.017861 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppss\" (UniqueName: \"kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.017949 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.017974 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.018071 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.057031 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.058499 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.063312 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.068836 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.096076 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.097697 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.099485 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.114580 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.120260 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.120374 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppss\" (UniqueName: \"kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.120432 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.120450 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.127426 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.135364 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.155822 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.158224 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppss\" (UniqueName: \"kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss\") pod \"nova-cell0-cell-mapping-ch8tx\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.217239 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224539 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224590 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfv8p\" (UniqueName: \"kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224704 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224723 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224741 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224821 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb588\" (UniqueName: \"kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.224860 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.236836 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.237487 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.238068 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.240182 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.243366 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.247522 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.259169 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.261416 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.278950 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.280857 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.304633 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326239 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96k76\" (UniqueName: \"kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326285 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326316 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326363 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326382 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfv8p\" (UniqueName: \"kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326405 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326422 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326439 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326454 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326501 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7x9p\" (UniqueName: \"kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326550 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326578 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.326597 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb588\" (UniqueName: \"kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.329074 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.334269 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.337077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.339474 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.343654 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.353645 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfv8p\" (UniqueName: \"kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.358815 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb588\" (UniqueName: \"kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588\") pod \"nova-metadata-0\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.375901 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.421054 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.434804 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.434889 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tttpn\" (UniqueName: \"kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.434940 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435019 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7x9p\" (UniqueName: \"kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435062 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435097 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435133 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435166 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435210 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435250 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96k76\" (UniqueName: \"kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435277 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.435327 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.438412 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.444633 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.444977 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.452822 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.453426 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.456045 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7x9p\" (UniqueName: \"kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p\") pod \"nova-api-0\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.459567 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96k76\" (UniqueName: \"kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76\") pod \"nova-scheduler-0\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.478706 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.537185 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.537374 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.537426 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.537463 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.537482 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tttpn\" (UniqueName: \"kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.538163 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.538766 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.540377 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.540379 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.570578 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tttpn\" (UniqueName: \"kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn\") pod \"dnsmasq-dns-74cc959f47-jbx4z\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.610977 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.787106 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.872820 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ch8tx"] Oct 14 08:52:17 crc kubenswrapper[5018]: I1014 08:52:17.987476 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ch8tx" event={"ID":"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2","Type":"ContainerStarted","Data":"6ddf69f27206dec20266f28b1289e2054415ffc83259aa1507eeaca2447faaea"} Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.039172 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.050197 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:52:18 crc kubenswrapper[5018]: W1014 08:52:18.060126 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7d3dcbc_0758_4a84_be9c_a33506064e75.slice/crio-f476977d0768af3c13f14ecbbead6d0d298cfc76505bdd676587bff3ea42521f WatchSource:0}: Error finding container f476977d0768af3c13f14ecbbead6d0d298cfc76505bdd676587bff3ea42521f: Status 404 returned error can't find the container with id f476977d0768af3c13f14ecbbead6d0d298cfc76505bdd676587bff3ea42521f Oct 14 08:52:18 crc kubenswrapper[5018]: W1014 08:52:18.061440 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddebf189c_b670_4639_8703_bd6656795416.slice/crio-03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7 WatchSource:0}: Error finding container 03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7: Status 404 returned error can't find the container with id 03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7 Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.061826 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.076502 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.108646 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8f5h4"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.110156 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.112999 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.113533 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.115539 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8f5h4"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.256553 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.256980 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.257049 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.257088 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptwwl\" (UniqueName: \"kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.300967 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.343821 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.362531 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.362592 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.362634 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.362655 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptwwl\" (UniqueName: \"kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.366705 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.370274 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.372184 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.383043 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptwwl\" (UniqueName: \"kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl\") pod \"nova-cell1-conductor-db-sync-8f5h4\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.429389 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:18 crc kubenswrapper[5018]: I1014 08:52:18.900832 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8f5h4"] Oct 14 08:52:18 crc kubenswrapper[5018]: W1014 08:52:18.908614 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90e54d5b_185d_40e9_a3a7_7c848eac44b1.slice/crio-396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9 WatchSource:0}: Error finding container 396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9: Status 404 returned error can't find the container with id 396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9 Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.002030 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ch8tx" event={"ID":"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2","Type":"ContainerStarted","Data":"ad2b4811966987ea30370b369c8a6b465e5f24ef16c13b9903cbbbc919d7aa05"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.005381 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerStarted","Data":"f476977d0768af3c13f14ecbbead6d0d298cfc76505bdd676587bff3ea42521f"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.009336 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2096278-58b2-4552-8221-00bfa330af8c","Type":"ContainerStarted","Data":"dde91a1379eb672defa7d465b1bd353ea8269858f8733f2a5679d4ecca5958d1"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.011232 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"debf189c-b670-4639-8703-bd6656795416","Type":"ContainerStarted","Data":"03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.012643 5018 generic.go:334] "Generic (PLEG): container finished" podID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerID="a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f" exitCode=0 Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.012701 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" event={"ID":"9b01f3e1-d593-41e5-b338-4b6c07a24f2a","Type":"ContainerDied","Data":"a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.012723 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" event={"ID":"9b01f3e1-d593-41e5-b338-4b6c07a24f2a","Type":"ContainerStarted","Data":"33ac0f5bf92c5f6fde0fde6c62ad606fbb6db1597651847ee79ca755e0f680b7"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.017902 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" event={"ID":"90e54d5b-185d-40e9-a3a7-7c848eac44b1","Type":"ContainerStarted","Data":"396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.021544 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerStarted","Data":"34eb2386a577675e4423a444dff2aaef0438b842eb30189bf2c2d99d2aa38111"} Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.024743 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ch8tx" podStartSLOduration=3.024728113 podStartE2EDuration="3.024728113s" podCreationTimestamp="2025-10-14 08:52:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:19.012922844 +0000 UTC m=+7355.596969471" watchObservedRunningTime="2025-10-14 08:52:19.024728113 +0000 UTC m=+7355.608774740" Oct 14 08:52:19 crc kubenswrapper[5018]: I1014 08:52:19.604845 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:52:19 crc kubenswrapper[5018]: E1014 08:52:19.605489 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:52:20 crc kubenswrapper[5018]: I1014 08:52:20.042404 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" event={"ID":"90e54d5b-185d-40e9-a3a7-7c848eac44b1","Type":"ContainerStarted","Data":"e470dea1a4e6ffec6409d94627e8b61dab6f61e9336b8490b9c80edaa9a40879"} Oct 14 08:52:20 crc kubenswrapper[5018]: I1014 08:52:20.065704 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" podStartSLOduration=2.065686386 podStartE2EDuration="2.065686386s" podCreationTimestamp="2025-10-14 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:20.058952822 +0000 UTC m=+7356.642999439" watchObservedRunningTime="2025-10-14 08:52:20.065686386 +0000 UTC m=+7356.649733013" Oct 14 08:52:21 crc kubenswrapper[5018]: I1014 08:52:21.004098 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:21 crc kubenswrapper[5018]: I1014 08:52:21.024463 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:21 crc kubenswrapper[5018]: I1014 08:52:21.056476 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" event={"ID":"9b01f3e1-d593-41e5-b338-4b6c07a24f2a","Type":"ContainerStarted","Data":"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c"} Oct 14 08:52:21 crc kubenswrapper[5018]: I1014 08:52:21.079548 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" podStartSLOduration=4.0795294 podStartE2EDuration="4.0795294s" podCreationTimestamp="2025-10-14 08:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:21.072276272 +0000 UTC m=+7357.656322899" watchObservedRunningTime="2025-10-14 08:52:21.0795294 +0000 UTC m=+7357.663576027" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.067646 5018 generic.go:334] "Generic (PLEG): container finished" podID="90e54d5b-185d-40e9-a3a7-7c848eac44b1" containerID="e470dea1a4e6ffec6409d94627e8b61dab6f61e9336b8490b9c80edaa9a40879" exitCode=0 Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.067729 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" event={"ID":"90e54d5b-185d-40e9-a3a7-7c848eac44b1","Type":"ContainerDied","Data":"e470dea1a4e6ffec6409d94627e8b61dab6f61e9336b8490b9c80edaa9a40879"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.071416 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerStarted","Data":"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.071457 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerStarted","Data":"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.093654 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerStarted","Data":"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.093699 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerStarted","Data":"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.093872 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-log" containerID="cri-o://fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" gracePeriod=30 Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.093987 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-metadata" containerID="cri-o://8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" gracePeriod=30 Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.113006 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2096278-58b2-4552-8221-00bfa330af8c","Type":"ContainerStarted","Data":"aec3f2c84fb3adc5c35d327057d61ae5a5bf0400836f01dfa71c059a152632d7"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.114453 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.574154006 podStartE2EDuration="5.114427828s" podCreationTimestamp="2025-10-14 08:52:17 +0000 UTC" firstStartedPulling="2025-10-14 08:52:18.290590079 +0000 UTC m=+7354.874636706" lastFinishedPulling="2025-10-14 08:52:20.830863901 +0000 UTC m=+7357.414910528" observedRunningTime="2025-10-14 08:52:22.111882325 +0000 UTC m=+7358.695928952" watchObservedRunningTime="2025-10-14 08:52:22.114427828 +0000 UTC m=+7358.698474475" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.128040 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="debf189c-b670-4639-8703-bd6656795416" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://df8465136d92b044fec3c7f76559ba6e5b0c147ecfd89e58a800587269160316" gracePeriod=30 Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.128393 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"debf189c-b670-4639-8703-bd6656795416","Type":"ContainerStarted","Data":"df8465136d92b044fec3c7f76559ba6e5b0c147ecfd89e58a800587269160316"} Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.128432 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.151602 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.383127383 podStartE2EDuration="5.151580605s" podCreationTimestamp="2025-10-14 08:52:17 +0000 UTC" firstStartedPulling="2025-10-14 08:52:18.067984639 +0000 UTC m=+7354.652031266" lastFinishedPulling="2025-10-14 08:52:20.836437861 +0000 UTC m=+7357.420484488" observedRunningTime="2025-10-14 08:52:22.132007923 +0000 UTC m=+7358.716054570" watchObservedRunningTime="2025-10-14 08:52:22.151580605 +0000 UTC m=+7358.735627232" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.159285 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.377024917 podStartE2EDuration="5.159274585s" podCreationTimestamp="2025-10-14 08:52:17 +0000 UTC" firstStartedPulling="2025-10-14 08:52:18.050005193 +0000 UTC m=+7354.634051820" lastFinishedPulling="2025-10-14 08:52:20.832254861 +0000 UTC m=+7357.416301488" observedRunningTime="2025-10-14 08:52:22.151942355 +0000 UTC m=+7358.735988992" watchObservedRunningTime="2025-10-14 08:52:22.159274585 +0000 UTC m=+7358.743321212" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.175586 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.410514318 podStartE2EDuration="5.175568463s" podCreationTimestamp="2025-10-14 08:52:17 +0000 UTC" firstStartedPulling="2025-10-14 08:52:18.064496309 +0000 UTC m=+7354.648542926" lastFinishedPulling="2025-10-14 08:52:20.829550444 +0000 UTC m=+7357.413597071" observedRunningTime="2025-10-14 08:52:22.164039062 +0000 UTC m=+7358.748085689" watchObservedRunningTime="2025-10-14 08:52:22.175568463 +0000 UTC m=+7358.759615090" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.377173 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.421867 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.421944 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.479966 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.816222 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.963372 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb588\" (UniqueName: \"kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588\") pod \"a7d3dcbc-0758-4a84-be9c-a33506064e75\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.963445 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs\") pod \"a7d3dcbc-0758-4a84-be9c-a33506064e75\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.963497 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data\") pod \"a7d3dcbc-0758-4a84-be9c-a33506064e75\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.963761 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle\") pod \"a7d3dcbc-0758-4a84-be9c-a33506064e75\" (UID: \"a7d3dcbc-0758-4a84-be9c-a33506064e75\") " Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.965763 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs" (OuterVolumeSpecName: "logs") pod "a7d3dcbc-0758-4a84-be9c-a33506064e75" (UID: "a7d3dcbc-0758-4a84-be9c-a33506064e75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:52:22 crc kubenswrapper[5018]: I1014 08:52:22.974177 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588" (OuterVolumeSpecName: "kube-api-access-zb588") pod "a7d3dcbc-0758-4a84-be9c-a33506064e75" (UID: "a7d3dcbc-0758-4a84-be9c-a33506064e75"). InnerVolumeSpecName "kube-api-access-zb588". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.015233 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7d3dcbc-0758-4a84-be9c-a33506064e75" (UID: "a7d3dcbc-0758-4a84-be9c-a33506064e75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.021840 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data" (OuterVolumeSpecName: "config-data") pod "a7d3dcbc-0758-4a84-be9c-a33506064e75" (UID: "a7d3dcbc-0758-4a84-be9c-a33506064e75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.066406 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.066458 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb588\" (UniqueName: \"kubernetes.io/projected/a7d3dcbc-0758-4a84-be9c-a33506064e75-kube-api-access-zb588\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.066478 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d3dcbc-0758-4a84-be9c-a33506064e75-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.066495 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d3dcbc-0758-4a84-be9c-a33506064e75-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.138957 5018 generic.go:334] "Generic (PLEG): container finished" podID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerID="8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" exitCode=0 Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139009 5018 generic.go:334] "Generic (PLEG): container finished" podID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerID="fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" exitCode=143 Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139071 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139089 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerDied","Data":"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8"} Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139131 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerDied","Data":"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120"} Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139152 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7d3dcbc-0758-4a84-be9c-a33506064e75","Type":"ContainerDied","Data":"f476977d0768af3c13f14ecbbead6d0d298cfc76505bdd676587bff3ea42521f"} Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.139178 5018 scope.go:117] "RemoveContainer" containerID="8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.144256 5018 generic.go:334] "Generic (PLEG): container finished" podID="e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" containerID="ad2b4811966987ea30370b369c8a6b465e5f24ef16c13b9903cbbbc919d7aa05" exitCode=0 Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.145655 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ch8tx" event={"ID":"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2","Type":"ContainerDied","Data":"ad2b4811966987ea30370b369c8a6b465e5f24ef16c13b9903cbbbc919d7aa05"} Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.185034 5018 scope.go:117] "RemoveContainer" containerID="fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.230925 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.239728 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.253878 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:23 crc kubenswrapper[5018]: E1014 08:52:23.254334 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-metadata" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.254356 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-metadata" Oct 14 08:52:23 crc kubenswrapper[5018]: E1014 08:52:23.254394 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-log" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.254406 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-log" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.254786 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-log" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.254856 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" containerName="nova-metadata-metadata" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.256365 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.257859 5018 scope.go:117] "RemoveContainer" containerID="8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.259115 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:52:23 crc kubenswrapper[5018]: E1014 08:52:23.260270 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8\": container with ID starting with 8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8 not found: ID does not exist" containerID="8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.260313 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8"} err="failed to get container status \"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8\": rpc error: code = NotFound desc = could not find container \"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8\": container with ID starting with 8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8 not found: ID does not exist" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.260352 5018 scope.go:117] "RemoveContainer" containerID="fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.260531 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:52:23 crc kubenswrapper[5018]: E1014 08:52:23.260746 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120\": container with ID starting with fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120 not found: ID does not exist" containerID="fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.260780 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120"} err="failed to get container status \"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120\": rpc error: code = NotFound desc = could not find container \"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120\": container with ID starting with fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120 not found: ID does not exist" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.260799 5018 scope.go:117] "RemoveContainer" containerID="8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.261154 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8"} err="failed to get container status \"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8\": rpc error: code = NotFound desc = could not find container \"8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8\": container with ID starting with 8bf1b63283a7a2d624a6dfc49a98d60acfee1acbfe35494fd12ffad6d00e57b8 not found: ID does not exist" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.261181 5018 scope.go:117] "RemoveContainer" containerID="fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.261449 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120"} err="failed to get container status \"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120\": rpc error: code = NotFound desc = could not find container \"fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120\": container with ID starting with fbf36175477995f172d639950c5c0f4d8dcf815bdcc65f983e460b71d0633120 not found: ID does not exist" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.265072 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.372399 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.372465 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.372851 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.372962 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.373099 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmsp6\" (UniqueName: \"kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.476034 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.476109 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.476172 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmsp6\" (UniqueName: \"kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.476379 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.476421 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.477176 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.481655 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.490188 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.490615 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.493989 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmsp6\" (UniqueName: \"kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6\") pod \"nova-metadata-0\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.592220 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.606500 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.678605 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptwwl\" (UniqueName: \"kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl\") pod \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.679238 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts\") pod \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.679298 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle\") pod \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.679366 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data\") pod \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\" (UID: \"90e54d5b-185d-40e9-a3a7-7c848eac44b1\") " Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.688530 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts" (OuterVolumeSpecName: "scripts") pod "90e54d5b-185d-40e9-a3a7-7c848eac44b1" (UID: "90e54d5b-185d-40e9-a3a7-7c848eac44b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.688954 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl" (OuterVolumeSpecName: "kube-api-access-ptwwl") pod "90e54d5b-185d-40e9-a3a7-7c848eac44b1" (UID: "90e54d5b-185d-40e9-a3a7-7c848eac44b1"). InnerVolumeSpecName "kube-api-access-ptwwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.748137 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90e54d5b-185d-40e9-a3a7-7c848eac44b1" (UID: "90e54d5b-185d-40e9-a3a7-7c848eac44b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.752786 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data" (OuterVolumeSpecName: "config-data") pod "90e54d5b-185d-40e9-a3a7-7c848eac44b1" (UID: "90e54d5b-185d-40e9-a3a7-7c848eac44b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.781937 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptwwl\" (UniqueName: \"kubernetes.io/projected/90e54d5b-185d-40e9-a3a7-7c848eac44b1-kube-api-access-ptwwl\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.781969 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.781983 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:23 crc kubenswrapper[5018]: I1014 08:52:23.781994 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e54d5b-185d-40e9-a3a7-7c848eac44b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.036315 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-v97s4"] Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.047689 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-v97s4"] Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.124065 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.153099 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.153154 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8f5h4" event={"ID":"90e54d5b-185d-40e9-a3a7-7c848eac44b1","Type":"ContainerDied","Data":"396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9"} Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.153198 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="396e8f38144a69cde8ad7af22dd2ff4269b9db0a342399718c07989c6850bba9" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.159922 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerStarted","Data":"1919e6c453dced2c12418d85d8332efb864eec9e7956082f6b8ac6ca1a56836a"} Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.221759 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:52:24 crc kubenswrapper[5018]: E1014 08:52:24.223018 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e54d5b-185d-40e9-a3a7-7c848eac44b1" containerName="nova-cell1-conductor-db-sync" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.223046 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e54d5b-185d-40e9-a3a7-7c848eac44b1" containerName="nova-cell1-conductor-db-sync" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.223503 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e54d5b-185d-40e9-a3a7-7c848eac44b1" containerName="nova-cell1-conductor-db-sync" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.224616 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.228441 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.236398 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.294064 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.294124 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.294180 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v64vt\" (UniqueName: \"kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.396150 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.396232 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.396829 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v64vt\" (UniqueName: \"kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.408040 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.408872 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.413765 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v64vt\" (UniqueName: \"kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt\") pod \"nova-cell1-conductor-0\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.519156 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.599201 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts\") pod \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.599788 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.599796 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data\") pod \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.599991 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qppss\" (UniqueName: \"kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss\") pod \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.600087 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle\") pod \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\" (UID: \"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2\") " Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.603240 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts" (OuterVolumeSpecName: "scripts") pod "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" (UID: "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.607798 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss" (OuterVolumeSpecName: "kube-api-access-qppss") pod "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" (UID: "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2"). InnerVolumeSpecName "kube-api-access-qppss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.622251 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d3dcbc-0758-4a84-be9c-a33506064e75" path="/var/lib/kubelet/pods/a7d3dcbc-0758-4a84-be9c-a33506064e75/volumes" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.622925 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7661bd-4acd-4bc1-9095-528af96485a9" path="/var/lib/kubelet/pods/ed7661bd-4acd-4bc1-9095-528af96485a9/volumes" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.626724 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" (UID: "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.642273 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data" (OuterVolumeSpecName: "config-data") pod "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" (UID: "e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.718238 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.718268 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.718281 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qppss\" (UniqueName: \"kubernetes.io/projected/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-kube-api-access-qppss\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:24 crc kubenswrapper[5018]: I1014 08:52:24.718290 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.108754 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:52:25 crc kubenswrapper[5018]: W1014 08:52:25.110207 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46a6263a_f328_43d0_8fcd_fb3610e88c30.slice/crio-e7ef6f4182de9825c89df33a03ee6d4c007d647ba422ba1d838a84bd08c311fe WatchSource:0}: Error finding container e7ef6f4182de9825c89df33a03ee6d4c007d647ba422ba1d838a84bd08c311fe: Status 404 returned error can't find the container with id e7ef6f4182de9825c89df33a03ee6d4c007d647ba422ba1d838a84bd08c311fe Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.186849 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ch8tx" event={"ID":"e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2","Type":"ContainerDied","Data":"6ddf69f27206dec20266f28b1289e2054415ffc83259aa1507eeaca2447faaea"} Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.186937 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ddf69f27206dec20266f28b1289e2054415ffc83259aa1507eeaca2447faaea" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.195155 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46a6263a-f328-43d0-8fcd-fb3610e88c30","Type":"ContainerStarted","Data":"e7ef6f4182de9825c89df33a03ee6d4c007d647ba422ba1d838a84bd08c311fe"} Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.200537 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerStarted","Data":"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34"} Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.200615 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerStarted","Data":"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2"} Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.201159 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ch8tx" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.242680 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.242655868 podStartE2EDuration="2.242655868s" podCreationTimestamp="2025-10-14 08:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:25.23856953 +0000 UTC m=+7361.822616157" watchObservedRunningTime="2025-10-14 08:52:25.242655868 +0000 UTC m=+7361.826702535" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.355688 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.355897 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-log" containerID="cri-o://3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" gracePeriod=30 Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.356309 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-api" containerID="cri-o://24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" gracePeriod=30 Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.406402 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.406672 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b2096278-58b2-4552-8221-00bfa330af8c" containerName="nova-scheduler-scheduler" containerID="cri-o://aec3f2c84fb3adc5c35d327057d61ae5a5bf0400836f01dfa71c059a152632d7" gracePeriod=30 Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.423687 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.876316 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.943920 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle\") pod \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.944163 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7x9p\" (UniqueName: \"kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p\") pod \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.944218 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs\") pod \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.944283 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data\") pod \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\" (UID: \"c4326e1b-87ed-429a-88aa-7f42dc053b8d\") " Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.945209 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs" (OuterVolumeSpecName: "logs") pod "c4326e1b-87ed-429a-88aa-7f42dc053b8d" (UID: "c4326e1b-87ed-429a-88aa-7f42dc053b8d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.948198 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p" (OuterVolumeSpecName: "kube-api-access-t7x9p") pod "c4326e1b-87ed-429a-88aa-7f42dc053b8d" (UID: "c4326e1b-87ed-429a-88aa-7f42dc053b8d"). InnerVolumeSpecName "kube-api-access-t7x9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.971421 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4326e1b-87ed-429a-88aa-7f42dc053b8d" (UID: "c4326e1b-87ed-429a-88aa-7f42dc053b8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[5018]: I1014 08:52:25.971453 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data" (OuterVolumeSpecName: "config-data") pod "c4326e1b-87ed-429a-88aa-7f42dc053b8d" (UID: "c4326e1b-87ed-429a-88aa-7f42dc053b8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.048976 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.049030 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7x9p\" (UniqueName: \"kubernetes.io/projected/c4326e1b-87ed-429a-88aa-7f42dc053b8d-kube-api-access-t7x9p\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.049056 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4326e1b-87ed-429a-88aa-7f42dc053b8d-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.049078 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4326e1b-87ed-429a-88aa-7f42dc053b8d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.212716 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46a6263a-f328-43d0-8fcd-fb3610e88c30","Type":"ContainerStarted","Data":"9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400"} Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.215045 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216727 5018 generic.go:334] "Generic (PLEG): container finished" podID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerID="24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" exitCode=0 Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216765 5018 generic.go:334] "Generic (PLEG): container finished" podID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerID="3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" exitCode=143 Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216788 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216809 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerDied","Data":"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d"} Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216827 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerDied","Data":"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697"} Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216839 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4326e1b-87ed-429a-88aa-7f42dc053b8d","Type":"ContainerDied","Data":"34eb2386a577675e4423a444dff2aaef0438b842eb30189bf2c2d99d2aa38111"} Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.216854 5018 scope.go:117] "RemoveContainer" containerID="24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.254582 5018 scope.go:117] "RemoveContainer" containerID="3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.257996 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.257951043 podStartE2EDuration="2.257951043s" podCreationTimestamp="2025-10-14 08:52:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:26.236157887 +0000 UTC m=+7362.820204554" watchObservedRunningTime="2025-10-14 08:52:26.257951043 +0000 UTC m=+7362.841997680" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.300649 5018 scope.go:117] "RemoveContainer" containerID="24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" Oct 14 08:52:26 crc kubenswrapper[5018]: E1014 08:52:26.301168 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d\": container with ID starting with 24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d not found: ID does not exist" containerID="24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.301195 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d"} err="failed to get container status \"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d\": rpc error: code = NotFound desc = could not find container \"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d\": container with ID starting with 24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d not found: ID does not exist" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.301222 5018 scope.go:117] "RemoveContainer" containerID="3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.301473 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.309733 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:26 crc kubenswrapper[5018]: E1014 08:52:26.318820 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697\": container with ID starting with 3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697 not found: ID does not exist" containerID="3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.318861 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697"} err="failed to get container status \"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697\": rpc error: code = NotFound desc = could not find container \"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697\": container with ID starting with 3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697 not found: ID does not exist" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.318889 5018 scope.go:117] "RemoveContainer" containerID="24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.322466 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d"} err="failed to get container status \"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d\": rpc error: code = NotFound desc = could not find container \"24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d\": container with ID starting with 24d3b66b14105ddee05a8f9cd1bdc2e342c3016e6ee3e51537c078d82388386d not found: ID does not exist" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.322518 5018 scope.go:117] "RemoveContainer" containerID="3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.322835 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697"} err="failed to get container status \"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697\": rpc error: code = NotFound desc = could not find container \"3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697\": container with ID starting with 3ffc3505937793de397e182267a14a6c838923bccc2ed5bdb51a54a39e5ac697 not found: ID does not exist" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.323720 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:26 crc kubenswrapper[5018]: E1014 08:52:26.324175 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" containerName="nova-manage" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324193 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" containerName="nova-manage" Oct 14 08:52:26 crc kubenswrapper[5018]: E1014 08:52:26.324206 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-log" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324212 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-log" Oct 14 08:52:26 crc kubenswrapper[5018]: E1014 08:52:26.324240 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-api" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324316 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-api" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324496 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-api" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324515 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" containerName="nova-api-log" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.324527 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" containerName="nova-manage" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.325545 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.329642 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.332807 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.457356 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.457544 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdd2j\" (UniqueName: \"kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.457614 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.457947 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.559514 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.559616 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.559677 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdd2j\" (UniqueName: \"kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.559703 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.560384 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.565129 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.566679 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.589073 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdd2j\" (UniqueName: \"kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j\") pod \"nova-api-0\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " pod="openstack/nova-api-0" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.618763 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4326e1b-87ed-429a-88aa-7f42dc053b8d" path="/var/lib/kubelet/pods/c4326e1b-87ed-429a-88aa-7f42dc053b8d/volumes" Oct 14 08:52:26 crc kubenswrapper[5018]: I1014 08:52:26.644719 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.186160 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:52:27 crc kubenswrapper[5018]: W1014 08:52:27.188488 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccd09ecc_4901_4232_b07b_b0e757534169.slice/crio-ea99b1141e6a47e05e49f1390dde300949bed9b05d87a441ddb65b4d0d34f827 WatchSource:0}: Error finding container ea99b1141e6a47e05e49f1390dde300949bed9b05d87a441ddb65b4d0d34f827: Status 404 returned error can't find the container with id ea99b1141e6a47e05e49f1390dde300949bed9b05d87a441ddb65b4d0d34f827 Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.228821 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerStarted","Data":"ea99b1141e6a47e05e49f1390dde300949bed9b05d87a441ddb65b4d0d34f827"} Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.229099 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-log" containerID="cri-o://482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" gracePeriod=30 Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.229151 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-metadata" containerID="cri-o://9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" gracePeriod=30 Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.731051 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.788815 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.797577 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle\") pod \"fe1a5a04-596e-4da9-ac44-e470bd78d368\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.798141 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs\") pod \"fe1a5a04-596e-4da9-ac44-e470bd78d368\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.798174 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs" (OuterVolumeSpecName: "logs") pod "fe1a5a04-596e-4da9-ac44-e470bd78d368" (UID: "fe1a5a04-596e-4da9-ac44-e470bd78d368"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.798260 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmsp6\" (UniqueName: \"kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6\") pod \"fe1a5a04-596e-4da9-ac44-e470bd78d368\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.798517 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs\") pod \"fe1a5a04-596e-4da9-ac44-e470bd78d368\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.798550 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data\") pod \"fe1a5a04-596e-4da9-ac44-e470bd78d368\" (UID: \"fe1a5a04-596e-4da9-ac44-e470bd78d368\") " Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.799287 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe1a5a04-596e-4da9-ac44-e470bd78d368-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.801920 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6" (OuterVolumeSpecName: "kube-api-access-gmsp6") pod "fe1a5a04-596e-4da9-ac44-e470bd78d368" (UID: "fe1a5a04-596e-4da9-ac44-e470bd78d368"). InnerVolumeSpecName "kube-api-access-gmsp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.848459 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.849085 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="dnsmasq-dns" containerID="cri-o://e7ce72802b638ad8af5e63ac212ad5735398a50811daa5c990e45f24450302f6" gracePeriod=10 Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.875182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data" (OuterVolumeSpecName: "config-data") pod "fe1a5a04-596e-4da9-ac44-e470bd78d368" (UID: "fe1a5a04-596e-4da9-ac44-e470bd78d368"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.885139 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe1a5a04-596e-4da9-ac44-e470bd78d368" (UID: "fe1a5a04-596e-4da9-ac44-e470bd78d368"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.901358 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.901395 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmsp6\" (UniqueName: \"kubernetes.io/projected/fe1a5a04-596e-4da9-ac44-e470bd78d368-kube-api-access-gmsp6\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.901410 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.910345 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fe1a5a04-596e-4da9-ac44-e470bd78d368" (UID: "fe1a5a04-596e-4da9-ac44-e470bd78d368"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:27 crc kubenswrapper[5018]: I1014 08:52:27.946629 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.82:5353: connect: connection refused" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.002737 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe1a5a04-596e-4da9-ac44-e470bd78d368-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247647 5018 generic.go:334] "Generic (PLEG): container finished" podID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerID="9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" exitCode=0 Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247677 5018 generic.go:334] "Generic (PLEG): container finished" podID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerID="482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" exitCode=143 Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247719 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerDied","Data":"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247747 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerDied","Data":"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247759 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe1a5a04-596e-4da9-ac44-e470bd78d368","Type":"ContainerDied","Data":"1919e6c453dced2c12418d85d8332efb864eec9e7956082f6b8ac6ca1a56836a"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247775 5018 scope.go:117] "RemoveContainer" containerID="9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.247906 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.255342 5018 generic.go:334] "Generic (PLEG): container finished" podID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerID="e7ce72802b638ad8af5e63ac212ad5735398a50811daa5c990e45f24450302f6" exitCode=0 Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.255412 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" event={"ID":"298fed88-8ba4-44fa-9327-ec27e95865b4","Type":"ContainerDied","Data":"e7ce72802b638ad8af5e63ac212ad5735398a50811daa5c990e45f24450302f6"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.255436 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" event={"ID":"298fed88-8ba4-44fa-9327-ec27e95865b4","Type":"ContainerDied","Data":"f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.255449 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f171be571415dc447a9be1c297f3a0edc672bea03faa2c0bbb3b14ff494ddf2d" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.257463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerStarted","Data":"551f2f1ba4fbe2a1887ac8b1f1cf84c5318bf6f2f1ab2a674db1cadaad562451"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.257510 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerStarted","Data":"a3187fe15e3254efe20b8520456943a4914d02f03ba058b9b83657e4fb750a87"} Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.278862 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2788488559999998 podStartE2EDuration="2.278848856s" podCreationTimestamp="2025-10-14 08:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:28.27622301 +0000 UTC m=+7364.860269647" watchObservedRunningTime="2025-10-14 08:52:28.278848856 +0000 UTC m=+7364.862895483" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.303231 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.314509 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.325390 5018 scope.go:117] "RemoveContainer" containerID="482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.338812 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349062 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.349555 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="dnsmasq-dns" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349571 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="dnsmasq-dns" Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.349592 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-metadata" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349603 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-metadata" Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.349632 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-log" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349642 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-log" Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.349667 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="init" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349674 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="init" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349763 5018 scope.go:117] "RemoveContainer" containerID="9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349842 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-log" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349862 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" containerName="dnsmasq-dns" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.349878 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" containerName="nova-metadata-metadata" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.350866 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.352225 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34\": container with ID starting with 9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34 not found: ID does not exist" containerID="9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.352257 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34"} err="failed to get container status \"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34\": rpc error: code = NotFound desc = could not find container \"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34\": container with ID starting with 9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34 not found: ID does not exist" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.352292 5018 scope.go:117] "RemoveContainer" containerID="482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" Oct 14 08:52:28 crc kubenswrapper[5018]: E1014 08:52:28.352943 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2\": container with ID starting with 482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2 not found: ID does not exist" containerID="482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.353007 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2"} err="failed to get container status \"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2\": rpc error: code = NotFound desc = could not find container \"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2\": container with ID starting with 482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2 not found: ID does not exist" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.353033 5018 scope.go:117] "RemoveContainer" containerID="9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.353147 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.353609 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.354593 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34"} err="failed to get container status \"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34\": rpc error: code = NotFound desc = could not find container \"9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34\": container with ID starting with 9e420476d8bca82645f5de8ee9dc378ea6e58faf3b780c47f96200aa6aee1c34 not found: ID does not exist" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.354628 5018 scope.go:117] "RemoveContainer" containerID="482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.359202 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2"} err="failed to get container status \"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2\": rpc error: code = NotFound desc = could not find container \"482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2\": container with ID starting with 482b541ab9b056f8763018dce42ef637ad3b1be39b0251550d5ccbcb442d59f2 not found: ID does not exist" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.366711 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.416982 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7fp7\" (UniqueName: \"kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7\") pod \"298fed88-8ba4-44fa-9327-ec27e95865b4\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417041 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb\") pod \"298fed88-8ba4-44fa-9327-ec27e95865b4\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417094 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb\") pod \"298fed88-8ba4-44fa-9327-ec27e95865b4\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417260 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc\") pod \"298fed88-8ba4-44fa-9327-ec27e95865b4\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417371 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config\") pod \"298fed88-8ba4-44fa-9327-ec27e95865b4\" (UID: \"298fed88-8ba4-44fa-9327-ec27e95865b4\") " Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417633 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq5jm\" (UniqueName: \"kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.417661 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.418430 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.418469 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.418667 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.422120 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7" (OuterVolumeSpecName: "kube-api-access-v7fp7") pod "298fed88-8ba4-44fa-9327-ec27e95865b4" (UID: "298fed88-8ba4-44fa-9327-ec27e95865b4"). InnerVolumeSpecName "kube-api-access-v7fp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.465804 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "298fed88-8ba4-44fa-9327-ec27e95865b4" (UID: "298fed88-8ba4-44fa-9327-ec27e95865b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.472235 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config" (OuterVolumeSpecName: "config") pod "298fed88-8ba4-44fa-9327-ec27e95865b4" (UID: "298fed88-8ba4-44fa-9327-ec27e95865b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.474656 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "298fed88-8ba4-44fa-9327-ec27e95865b4" (UID: "298fed88-8ba4-44fa-9327-ec27e95865b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.478126 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "298fed88-8ba4-44fa-9327-ec27e95865b4" (UID: "298fed88-8ba4-44fa-9327-ec27e95865b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520145 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520186 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520280 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520301 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq5jm\" (UniqueName: \"kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520319 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520368 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520378 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520388 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7fp7\" (UniqueName: \"kubernetes.io/projected/298fed88-8ba4-44fa-9327-ec27e95865b4-kube-api-access-v7fp7\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520398 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520407 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/298fed88-8ba4-44fa-9327-ec27e95865b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.520712 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.524408 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.524507 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.524997 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.539365 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq5jm\" (UniqueName: \"kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm\") pod \"nova-metadata-0\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " pod="openstack/nova-metadata-0" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.617669 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe1a5a04-596e-4da9-ac44-e470bd78d368" path="/var/lib/kubelet/pods/fe1a5a04-596e-4da9-ac44-e470bd78d368/volumes" Oct 14 08:52:28 crc kubenswrapper[5018]: I1014 08:52:28.663549 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:52:29 crc kubenswrapper[5018]: I1014 08:52:29.136839 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:52:29 crc kubenswrapper[5018]: W1014 08:52:29.137472 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46763a04_7b4c_45c6_9448_81a8776226d7.slice/crio-cd60b67d07b77a5595444828e0ff0303845d3ec940eafbde31cbb77ae115afdb WatchSource:0}: Error finding container cd60b67d07b77a5595444828e0ff0303845d3ec940eafbde31cbb77ae115afdb: Status 404 returned error can't find the container with id cd60b67d07b77a5595444828e0ff0303845d3ec940eafbde31cbb77ae115afdb Oct 14 08:52:29 crc kubenswrapper[5018]: I1014 08:52:29.282403 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerStarted","Data":"cd60b67d07b77a5595444828e0ff0303845d3ec940eafbde31cbb77ae115afdb"} Oct 14 08:52:29 crc kubenswrapper[5018]: I1014 08:52:29.282420 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9759f875c-7kzp6" Oct 14 08:52:29 crc kubenswrapper[5018]: I1014 08:52:29.313075 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:52:29 crc kubenswrapper[5018]: I1014 08:52:29.326527 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9759f875c-7kzp6"] Oct 14 08:52:30 crc kubenswrapper[5018]: I1014 08:52:30.297663 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerStarted","Data":"d542da88580180945500c895d35d137a7d6986b82d12ff9e7e5efeb1e7dc245a"} Oct 14 08:52:30 crc kubenswrapper[5018]: I1014 08:52:30.298011 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerStarted","Data":"168d82fe17a289b220662f688e9b409e15b6cebcdc4cec8740a5e060d7b29738"} Oct 14 08:52:30 crc kubenswrapper[5018]: I1014 08:52:30.327227 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.327196866 podStartE2EDuration="2.327196866s" podCreationTimestamp="2025-10-14 08:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:30.326485466 +0000 UTC m=+7366.910532163" watchObservedRunningTime="2025-10-14 08:52:30.327196866 +0000 UTC m=+7366.911243563" Oct 14 08:52:30 crc kubenswrapper[5018]: I1014 08:52:30.623707 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298fed88-8ba4-44fa-9327-ec27e95865b4" path="/var/lib/kubelet/pods/298fed88-8ba4-44fa-9327-ec27e95865b4/volumes" Oct 14 08:52:33 crc kubenswrapper[5018]: I1014 08:52:33.605880 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:52:33 crc kubenswrapper[5018]: I1014 08:52:33.663810 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:52:33 crc kubenswrapper[5018]: I1014 08:52:33.663906 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:52:34 crc kubenswrapper[5018]: I1014 08:52:34.339225 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e"} Oct 14 08:52:34 crc kubenswrapper[5018]: I1014 08:52:34.647760 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 08:52:36 crc kubenswrapper[5018]: I1014 08:52:36.645013 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:52:36 crc kubenswrapper[5018]: I1014 08:52:36.646006 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:52:37 crc kubenswrapper[5018]: I1014 08:52:37.052434 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6wjb4"] Oct 14 08:52:37 crc kubenswrapper[5018]: I1014 08:52:37.062733 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6wjb4"] Oct 14 08:52:37 crc kubenswrapper[5018]: I1014 08:52:37.727029 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:37 crc kubenswrapper[5018]: I1014 08:52:37.727437 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:38 crc kubenswrapper[5018]: I1014 08:52:38.628132 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ddd357a-32f5-4a4a-954a-50740f3d4e23" path="/var/lib/kubelet/pods/3ddd357a-32f5-4a4a-954a-50740f3d4e23/volumes" Oct 14 08:52:38 crc kubenswrapper[5018]: I1014 08:52:38.664746 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:52:38 crc kubenswrapper[5018]: I1014 08:52:38.664883 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:52:39 crc kubenswrapper[5018]: I1014 08:52:39.686886 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:39 crc kubenswrapper[5018]: I1014 08:52:39.686991 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:47 crc kubenswrapper[5018]: I1014 08:52:47.727904 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:47 crc kubenswrapper[5018]: I1014 08:52:47.727906 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:49 crc kubenswrapper[5018]: I1014 08:52:49.672747 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:49 crc kubenswrapper[5018]: I1014 08:52:49.672821 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.582366 5018 generic.go:334] "Generic (PLEG): container finished" podID="debf189c-b670-4639-8703-bd6656795416" containerID="df8465136d92b044fec3c7f76559ba6e5b0c147ecfd89e58a800587269160316" exitCode=137 Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.583362 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"debf189c-b670-4639-8703-bd6656795416","Type":"ContainerDied","Data":"df8465136d92b044fec3c7f76559ba6e5b0c147ecfd89e58a800587269160316"} Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.583527 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"debf189c-b670-4639-8703-bd6656795416","Type":"ContainerDied","Data":"03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7"} Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.583553 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03236160ec8808c89492411c86b6605ac408de9bed309f974629c0522104e3a7" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.633400 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.769864 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle\") pod \"debf189c-b670-4639-8703-bd6656795416\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.770408 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfv8p\" (UniqueName: \"kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p\") pod \"debf189c-b670-4639-8703-bd6656795416\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.770689 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data\") pod \"debf189c-b670-4639-8703-bd6656795416\" (UID: \"debf189c-b670-4639-8703-bd6656795416\") " Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.776801 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p" (OuterVolumeSpecName: "kube-api-access-lfv8p") pod "debf189c-b670-4639-8703-bd6656795416" (UID: "debf189c-b670-4639-8703-bd6656795416"). InnerVolumeSpecName "kube-api-access-lfv8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.820436 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "debf189c-b670-4639-8703-bd6656795416" (UID: "debf189c-b670-4639-8703-bd6656795416"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.820981 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data" (OuterVolumeSpecName: "config-data") pod "debf189c-b670-4639-8703-bd6656795416" (UID: "debf189c-b670-4639-8703-bd6656795416"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.873674 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.873719 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/debf189c-b670-4639-8703-bd6656795416-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:52 crc kubenswrapper[5018]: I1014 08:52:52.873734 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfv8p\" (UniqueName: \"kubernetes.io/projected/debf189c-b670-4639-8703-bd6656795416-kube-api-access-lfv8p\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.598046 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.652300 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.661308 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.687560 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:53 crc kubenswrapper[5018]: E1014 08:52:53.688081 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debf189c-b670-4639-8703-bd6656795416" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.688103 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="debf189c-b670-4639-8703-bd6656795416" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.688305 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="debf189c-b670-4639-8703-bd6656795416" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.689043 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.694647 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.694965 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.701856 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.722957 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.800862 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.801181 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.801592 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.801756 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6g2g\" (UniqueName: \"kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.801911 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.903755 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.903843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6g2g\" (UniqueName: \"kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.903929 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.903975 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.904175 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.911231 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.911657 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.913992 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.915559 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:53 crc kubenswrapper[5018]: I1014 08:52:53.929891 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6g2g\" (UniqueName: \"kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g\") pod \"nova-cell1-novncproxy-0\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:54 crc kubenswrapper[5018]: I1014 08:52:54.024442 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:54 crc kubenswrapper[5018]: I1014 08:52:54.505411 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:52:54 crc kubenswrapper[5018]: W1014 08:52:54.508117 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9830a080_9886_4018_b68c_60bda6dd6333.slice/crio-df3d3f9450e3982d493280cbc646c2305b11a691c73e595e30de2b0b3307f2d4 WatchSource:0}: Error finding container df3d3f9450e3982d493280cbc646c2305b11a691c73e595e30de2b0b3307f2d4: Status 404 returned error can't find the container with id df3d3f9450e3982d493280cbc646c2305b11a691c73e595e30de2b0b3307f2d4 Oct 14 08:52:54 crc kubenswrapper[5018]: I1014 08:52:54.618754 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="debf189c-b670-4639-8703-bd6656795416" path="/var/lib/kubelet/pods/debf189c-b670-4639-8703-bd6656795416/volumes" Oct 14 08:52:54 crc kubenswrapper[5018]: I1014 08:52:54.619909 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9830a080-9886-4018-b68c-60bda6dd6333","Type":"ContainerStarted","Data":"df3d3f9450e3982d493280cbc646c2305b11a691c73e595e30de2b0b3307f2d4"} Oct 14 08:52:55 crc kubenswrapper[5018]: I1014 08:52:55.626391 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9830a080-9886-4018-b68c-60bda6dd6333","Type":"ContainerStarted","Data":"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb"} Oct 14 08:52:55 crc kubenswrapper[5018]: I1014 08:52:55.633811 5018 generic.go:334] "Generic (PLEG): container finished" podID="b2096278-58b2-4552-8221-00bfa330af8c" containerID="aec3f2c84fb3adc5c35d327057d61ae5a5bf0400836f01dfa71c059a152632d7" exitCode=137 Oct 14 08:52:55 crc kubenswrapper[5018]: I1014 08:52:55.633861 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2096278-58b2-4552-8221-00bfa330af8c","Type":"ContainerDied","Data":"aec3f2c84fb3adc5c35d327057d61ae5a5bf0400836f01dfa71c059a152632d7"} Oct 14 08:52:55 crc kubenswrapper[5018]: I1014 08:52:55.665731 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.665706537 podStartE2EDuration="2.665706537s" podCreationTimestamp="2025-10-14 08:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:55.656951605 +0000 UTC m=+7392.240998242" watchObservedRunningTime="2025-10-14 08:52:55.665706537 +0000 UTC m=+7392.249753174" Oct 14 08:52:55 crc kubenswrapper[5018]: I1014 08:52:55.980416 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.158345 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96k76\" (UniqueName: \"kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76\") pod \"b2096278-58b2-4552-8221-00bfa330af8c\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.158700 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data\") pod \"b2096278-58b2-4552-8221-00bfa330af8c\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.158985 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle\") pod \"b2096278-58b2-4552-8221-00bfa330af8c\" (UID: \"b2096278-58b2-4552-8221-00bfa330af8c\") " Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.178317 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76" (OuterVolumeSpecName: "kube-api-access-96k76") pod "b2096278-58b2-4552-8221-00bfa330af8c" (UID: "b2096278-58b2-4552-8221-00bfa330af8c"). InnerVolumeSpecName "kube-api-access-96k76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.212820 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data" (OuterVolumeSpecName: "config-data") pod "b2096278-58b2-4552-8221-00bfa330af8c" (UID: "b2096278-58b2-4552-8221-00bfa330af8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.220693 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2096278-58b2-4552-8221-00bfa330af8c" (UID: "b2096278-58b2-4552-8221-00bfa330af8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.262059 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.262120 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96k76\" (UniqueName: \"kubernetes.io/projected/b2096278-58b2-4552-8221-00bfa330af8c-kube-api-access-96k76\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.262143 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2096278-58b2-4552-8221-00bfa330af8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.644861 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.645052 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.646741 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b2096278-58b2-4552-8221-00bfa330af8c","Type":"ContainerDied","Data":"dde91a1379eb672defa7d465b1bd353ea8269858f8733f2a5679d4ecca5958d1"} Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.646821 5018 scope.go:117] "RemoveContainer" containerID="aec3f2c84fb3adc5c35d327057d61ae5a5bf0400836f01dfa71c059a152632d7" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.647102 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.681688 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.700710 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.713745 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:56 crc kubenswrapper[5018]: E1014 08:52:56.714581 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2096278-58b2-4552-8221-00bfa330af8c" containerName="nova-scheduler-scheduler" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.714840 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2096278-58b2-4552-8221-00bfa330af8c" containerName="nova-scheduler-scheduler" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.715378 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2096278-58b2-4552-8221-00bfa330af8c" containerName="nova-scheduler-scheduler" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.716792 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.720139 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.739449 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.874431 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.874595 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.874807 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ngqq\" (UniqueName: \"kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.976989 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ngqq\" (UniqueName: \"kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.977142 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.977179 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.982191 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:56 crc kubenswrapper[5018]: I1014 08:52:56.985206 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.007474 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ngqq\" (UniqueName: \"kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq\") pod \"nova-scheduler-0\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " pod="openstack/nova-scheduler-0" Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.045017 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:52:57 crc kubenswrapper[5018]: W1014 08:52:57.516997 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d87f8bc_88ae_442e_b342_1c888f199a53.slice/crio-34955393d349aee7ed2df928d1c500f78a222cf5176b1874939b75e5f4db92fe WatchSource:0}: Error finding container 34955393d349aee7ed2df928d1c500f78a222cf5176b1874939b75e5f4db92fe: Status 404 returned error can't find the container with id 34955393d349aee7ed2df928d1c500f78a222cf5176b1874939b75e5f4db92fe Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.518919 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.661058 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d87f8bc-88ae-442e-b342-1c888f199a53","Type":"ContainerStarted","Data":"34955393d349aee7ed2df928d1c500f78a222cf5176b1874939b75e5f4db92fe"} Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.686900 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:57 crc kubenswrapper[5018]: I1014 08:52:57.727874 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:58 crc kubenswrapper[5018]: I1014 08:52:58.623841 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2096278-58b2-4552-8221-00bfa330af8c" path="/var/lib/kubelet/pods/b2096278-58b2-4552-8221-00bfa330af8c/volumes" Oct 14 08:52:58 crc kubenswrapper[5018]: I1014 08:52:58.676725 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d87f8bc-88ae-442e-b342-1c888f199a53","Type":"ContainerStarted","Data":"2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f"} Oct 14 08:52:58 crc kubenswrapper[5018]: I1014 08:52:58.716465 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.716440042 podStartE2EDuration="2.716440042s" podCreationTimestamp="2025-10-14 08:52:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:58.700643259 +0000 UTC m=+7395.284689886" watchObservedRunningTime="2025-10-14 08:52:58.716440042 +0000 UTC m=+7395.300486679" Oct 14 08:52:59 crc kubenswrapper[5018]: I1014 08:52:59.025004 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:52:59 crc kubenswrapper[5018]: I1014 08:52:59.702875 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:52:59 crc kubenswrapper[5018]: I1014 08:52:59.702925 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:02 crc kubenswrapper[5018]: I1014 08:53:02.045861 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:53:04 crc kubenswrapper[5018]: I1014 08:53:04.025973 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:53:04 crc kubenswrapper[5018]: I1014 08:53:04.061360 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:53:04 crc kubenswrapper[5018]: I1014 08:53:04.808280 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.023670 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-xxvbb"] Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.025093 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.028263 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.028393 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.040411 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xxvbb"] Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.062090 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.062145 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.062240 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pd89\" (UniqueName: \"kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.062288 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.163845 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.163899 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.164010 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pd89\" (UniqueName: \"kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.164067 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.172166 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.175256 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.175922 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.183667 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pd89\" (UniqueName: \"kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89\") pod \"nova-cell1-cell-mapping-xxvbb\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:05 crc kubenswrapper[5018]: I1014 08:53:05.344765 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:06 crc kubenswrapper[5018]: I1014 08:53:05.931717 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xxvbb"] Oct 14 08:53:06 crc kubenswrapper[5018]: I1014 08:53:06.798275 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xxvbb" event={"ID":"68d71d68-03a3-46e4-98be-ebd0904194d2","Type":"ContainerStarted","Data":"43bfd079138bb3c2e9ec5326544cf4acf7e96fc023b1eadc3c147d3b1433cbb2"} Oct 14 08:53:06 crc kubenswrapper[5018]: I1014 08:53:06.798860 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xxvbb" event={"ID":"68d71d68-03a3-46e4-98be-ebd0904194d2","Type":"ContainerStarted","Data":"a6beb16a637486dbc2095122442bcdbc799d4900cbb7ceb5941783031b67fe56"} Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.046155 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.097439 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.130010 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-xxvbb" podStartSLOduration=2.129986595 podStartE2EDuration="2.129986595s" podCreationTimestamp="2025-10-14 08:53:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:06.837139788 +0000 UTC m=+7403.421186445" watchObservedRunningTime="2025-10-14 08:53:07.129986595 +0000 UTC m=+7403.714033232" Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.726922 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.727264 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.102:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:07 crc kubenswrapper[5018]: I1014 08:53:07.875883 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 08:53:09 crc kubenswrapper[5018]: I1014 08:53:09.064533 5018 scope.go:117] "RemoveContainer" containerID="736a9779931d58dc29fa4c031e5b54d1b6eb2ad4fce588ed956b779339e83771" Oct 14 08:53:09 crc kubenswrapper[5018]: I1014 08:53:09.115290 5018 scope.go:117] "RemoveContainer" containerID="24db27ccf8374e2c2a417d5c951352e914ac93536c05b816a358776448c6f0bb" Oct 14 08:53:09 crc kubenswrapper[5018]: I1014 08:53:09.162565 5018 scope.go:117] "RemoveContainer" containerID="a4f64eef94cfa3ab48e1dc2b370ab0b6f9daff4ffcbb1eceb12fadffc3f2ea61" Oct 14 08:53:09 crc kubenswrapper[5018]: I1014 08:53:09.672933 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:09 crc kubenswrapper[5018]: I1014 08:53:09.672942 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:10 crc kubenswrapper[5018]: I1014 08:53:10.847451 5018 generic.go:334] "Generic (PLEG): container finished" podID="68d71d68-03a3-46e4-98be-ebd0904194d2" containerID="43bfd079138bb3c2e9ec5326544cf4acf7e96fc023b1eadc3c147d3b1433cbb2" exitCode=0 Oct 14 08:53:10 crc kubenswrapper[5018]: I1014 08:53:10.847542 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xxvbb" event={"ID":"68d71d68-03a3-46e4-98be-ebd0904194d2","Type":"ContainerDied","Data":"43bfd079138bb3c2e9ec5326544cf4acf7e96fc023b1eadc3c147d3b1433cbb2"} Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.290403 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.447597 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pd89\" (UniqueName: \"kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89\") pod \"68d71d68-03a3-46e4-98be-ebd0904194d2\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.447665 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data\") pod \"68d71d68-03a3-46e4-98be-ebd0904194d2\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.447886 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts\") pod \"68d71d68-03a3-46e4-98be-ebd0904194d2\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.447915 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle\") pod \"68d71d68-03a3-46e4-98be-ebd0904194d2\" (UID: \"68d71d68-03a3-46e4-98be-ebd0904194d2\") " Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.460779 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts" (OuterVolumeSpecName: "scripts") pod "68d71d68-03a3-46e4-98be-ebd0904194d2" (UID: "68d71d68-03a3-46e4-98be-ebd0904194d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.460788 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89" (OuterVolumeSpecName: "kube-api-access-9pd89") pod "68d71d68-03a3-46e4-98be-ebd0904194d2" (UID: "68d71d68-03a3-46e4-98be-ebd0904194d2"). InnerVolumeSpecName "kube-api-access-9pd89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.488221 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68d71d68-03a3-46e4-98be-ebd0904194d2" (UID: "68d71d68-03a3-46e4-98be-ebd0904194d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.492824 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data" (OuterVolumeSpecName: "config-data") pod "68d71d68-03a3-46e4-98be-ebd0904194d2" (UID: "68d71d68-03a3-46e4-98be-ebd0904194d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.550542 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.550581 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.550594 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d71d68-03a3-46e4-98be-ebd0904194d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.550607 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pd89\" (UniqueName: \"kubernetes.io/projected/68d71d68-03a3-46e4-98be-ebd0904194d2-kube-api-access-9pd89\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.870604 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xxvbb" event={"ID":"68d71d68-03a3-46e4-98be-ebd0904194d2","Type":"ContainerDied","Data":"a6beb16a637486dbc2095122442bcdbc799d4900cbb7ceb5941783031b67fe56"} Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.870669 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6beb16a637486dbc2095122442bcdbc799d4900cbb7ceb5941783031b67fe56" Oct 14 08:53:12 crc kubenswrapper[5018]: I1014 08:53:12.871111 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xxvbb" Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.120699 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.120948 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" containerID="cri-o://2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" gracePeriod=30 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.139133 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.140126 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" containerID="cri-o://a3187fe15e3254efe20b8520456943a4914d02f03ba058b9b83657e4fb750a87" gracePeriod=30 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.140224 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" containerID="cri-o://551f2f1ba4fbe2a1887ac8b1f1cf84c5318bf6f2f1ab2a674db1cadaad562451" gracePeriod=30 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.148659 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.148875 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" containerID="cri-o://168d82fe17a289b220662f688e9b409e15b6cebcdc4cec8740a5e060d7b29738" gracePeriod=30 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.148973 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" containerID="cri-o://d542da88580180945500c895d35d137a7d6986b82d12ff9e7e5efeb1e7dc245a" gracePeriod=30 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.879171 5018 generic.go:334] "Generic (PLEG): container finished" podID="46763a04-7b4c-45c6-9448-81a8776226d7" containerID="168d82fe17a289b220662f688e9b409e15b6cebcdc4cec8740a5e060d7b29738" exitCode=143 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.879259 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerDied","Data":"168d82fe17a289b220662f688e9b409e15b6cebcdc4cec8740a5e060d7b29738"} Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.881317 5018 generic.go:334] "Generic (PLEG): container finished" podID="ccd09ecc-4901-4232-b07b-b0e757534169" containerID="a3187fe15e3254efe20b8520456943a4914d02f03ba058b9b83657e4fb750a87" exitCode=143 Oct 14 08:53:13 crc kubenswrapper[5018]: I1014 08:53:13.881385 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerDied","Data":"a3187fe15e3254efe20b8520456943a4914d02f03ba058b9b83657e4fb750a87"} Oct 14 08:53:17 crc kubenswrapper[5018]: E1014 08:53:17.048177 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:17 crc kubenswrapper[5018]: E1014 08:53:17.050527 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:17 crc kubenswrapper[5018]: E1014 08:53:17.055796 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:17 crc kubenswrapper[5018]: E1014 08:53:17.055842 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:22 crc kubenswrapper[5018]: E1014 08:53:22.048109 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:22 crc kubenswrapper[5018]: E1014 08:53:22.050514 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:22 crc kubenswrapper[5018]: E1014 08:53:22.052259 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:22 crc kubenswrapper[5018]: E1014 08:53:22.052296 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.032314 5018 generic.go:334] "Generic (PLEG): container finished" podID="ccd09ecc-4901-4232-b07b-b0e757534169" containerID="551f2f1ba4fbe2a1887ac8b1f1cf84c5318bf6f2f1ab2a674db1cadaad562451" exitCode=0 Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.034233 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerDied","Data":"551f2f1ba4fbe2a1887ac8b1f1cf84c5318bf6f2f1ab2a674db1cadaad562451"} Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.037621 5018 generic.go:334] "Generic (PLEG): container finished" podID="46763a04-7b4c-45c6-9448-81a8776226d7" containerID="d542da88580180945500c895d35d137a7d6986b82d12ff9e7e5efeb1e7dc245a" exitCode=0 Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.037649 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerDied","Data":"d542da88580180945500c895d35d137a7d6986b82d12ff9e7e5efeb1e7dc245a"} Oct 14 08:53:27 crc kubenswrapper[5018]: E1014 08:53:27.047856 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:27 crc kubenswrapper[5018]: E1014 08:53:27.052041 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:27 crc kubenswrapper[5018]: E1014 08:53:27.055114 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:27 crc kubenswrapper[5018]: E1014 08:53:27.055155 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.168103 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.173871 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272248 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq5jm\" (UniqueName: \"kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm\") pod \"46763a04-7b4c-45c6-9448-81a8776226d7\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272289 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle\") pod \"ccd09ecc-4901-4232-b07b-b0e757534169\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272324 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs\") pod \"ccd09ecc-4901-4232-b07b-b0e757534169\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs\") pod \"46763a04-7b4c-45c6-9448-81a8776226d7\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272380 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs\") pod \"46763a04-7b4c-45c6-9448-81a8776226d7\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272400 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdd2j\" (UniqueName: \"kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j\") pod \"ccd09ecc-4901-4232-b07b-b0e757534169\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272497 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle\") pod \"46763a04-7b4c-45c6-9448-81a8776226d7\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.272934 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs" (OuterVolumeSpecName: "logs") pod "46763a04-7b4c-45c6-9448-81a8776226d7" (UID: "46763a04-7b4c-45c6-9448-81a8776226d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.273145 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data\") pod \"46763a04-7b4c-45c6-9448-81a8776226d7\" (UID: \"46763a04-7b4c-45c6-9448-81a8776226d7\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.273165 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data\") pod \"ccd09ecc-4901-4232-b07b-b0e757534169\" (UID: \"ccd09ecc-4901-4232-b07b-b0e757534169\") " Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.273478 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs" (OuterVolumeSpecName: "logs") pod "ccd09ecc-4901-4232-b07b-b0e757534169" (UID: "ccd09ecc-4901-4232-b07b-b0e757534169"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.273590 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccd09ecc-4901-4232-b07b-b0e757534169-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.273606 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46763a04-7b4c-45c6-9448-81a8776226d7-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.278427 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm" (OuterVolumeSpecName: "kube-api-access-zq5jm") pod "46763a04-7b4c-45c6-9448-81a8776226d7" (UID: "46763a04-7b4c-45c6-9448-81a8776226d7"). InnerVolumeSpecName "kube-api-access-zq5jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.279177 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j" (OuterVolumeSpecName: "kube-api-access-tdd2j") pod "ccd09ecc-4901-4232-b07b-b0e757534169" (UID: "ccd09ecc-4901-4232-b07b-b0e757534169"). InnerVolumeSpecName "kube-api-access-tdd2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.301795 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccd09ecc-4901-4232-b07b-b0e757534169" (UID: "ccd09ecc-4901-4232-b07b-b0e757534169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.307511 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46763a04-7b4c-45c6-9448-81a8776226d7" (UID: "46763a04-7b4c-45c6-9448-81a8776226d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.315030 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data" (OuterVolumeSpecName: "config-data") pod "ccd09ecc-4901-4232-b07b-b0e757534169" (UID: "ccd09ecc-4901-4232-b07b-b0e757534169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.321503 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data" (OuterVolumeSpecName: "config-data") pod "46763a04-7b4c-45c6-9448-81a8776226d7" (UID: "46763a04-7b4c-45c6-9448-81a8776226d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.325977 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "46763a04-7b4c-45c6-9448-81a8776226d7" (UID: "46763a04-7b4c-45c6-9448-81a8776226d7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375667 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdd2j\" (UniqueName: \"kubernetes.io/projected/ccd09ecc-4901-4232-b07b-b0e757534169-kube-api-access-tdd2j\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375697 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375707 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375716 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375725 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq5jm\" (UniqueName: \"kubernetes.io/projected/46763a04-7b4c-45c6-9448-81a8776226d7-kube-api-access-zq5jm\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375733 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd09ecc-4901-4232-b07b-b0e757534169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:27 crc kubenswrapper[5018]: I1014 08:53:27.375740 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46763a04-7b4c-45c6-9448-81a8776226d7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.054718 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ccd09ecc-4901-4232-b07b-b0e757534169","Type":"ContainerDied","Data":"ea99b1141e6a47e05e49f1390dde300949bed9b05d87a441ddb65b4d0d34f827"} Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.055283 5018 scope.go:117] "RemoveContainer" containerID="551f2f1ba4fbe2a1887ac8b1f1cf84c5318bf6f2f1ab2a674db1cadaad562451" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.054764 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.059113 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46763a04-7b4c-45c6-9448-81a8776226d7","Type":"ContainerDied","Data":"cd60b67d07b77a5595444828e0ff0303845d3ec940eafbde31cbb77ae115afdb"} Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.059240 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.093849 5018 scope.go:117] "RemoveContainer" containerID="a3187fe15e3254efe20b8520456943a4914d02f03ba058b9b83657e4fb750a87" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.134365 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.136355 5018 scope.go:117] "RemoveContainer" containerID="d542da88580180945500c895d35d137a7d6986b82d12ff9e7e5efeb1e7dc245a" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.154888 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.172352 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.183169 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.191860 5018 scope.go:117] "RemoveContainer" containerID="168d82fe17a289b220662f688e9b409e15b6cebcdc4cec8740a5e060d7b29738" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.194430 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: E1014 08:53:28.194940 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d71d68-03a3-46e4-98be-ebd0904194d2" containerName="nova-manage" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.194964 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d71d68-03a3-46e4-98be-ebd0904194d2" containerName="nova-manage" Oct 14 08:53:28 crc kubenswrapper[5018]: E1014 08:53:28.194985 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.194994 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" Oct 14 08:53:28 crc kubenswrapper[5018]: E1014 08:53:28.195017 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195024 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" Oct 14 08:53:28 crc kubenswrapper[5018]: E1014 08:53:28.195048 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195055 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" Oct 14 08:53:28 crc kubenswrapper[5018]: E1014 08:53:28.195066 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195074 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195295 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-log" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195318 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" containerName="nova-api-api" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195338 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d71d68-03a3-46e4-98be-ebd0904194d2" containerName="nova-manage" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195349 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-log" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.195364 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" containerName="nova-metadata-metadata" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.196641 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.198297 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.205016 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.207799 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.209121 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.209878 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.217349 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.229155 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295474 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295534 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295562 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295619 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295727 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhd9f\" (UniqueName: \"kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295780 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295804 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295862 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.295908 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-262sj\" (UniqueName: \"kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398497 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398614 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398727 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398813 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-262sj\" (UniqueName: \"kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398917 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.398982 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.399034 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.399062 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.399173 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.399327 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhd9f\" (UniqueName: \"kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.400899 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.406325 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.406714 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.408017 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.408222 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.416459 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.429105 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-262sj\" (UniqueName: \"kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj\") pod \"nova-api-0\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.433553 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhd9f\" (UniqueName: \"kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f\") pod \"nova-metadata-0\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.549421 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.558079 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.623970 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46763a04-7b4c-45c6-9448-81a8776226d7" path="/var/lib/kubelet/pods/46763a04-7b4c-45c6-9448-81a8776226d7/volumes" Oct 14 08:53:28 crc kubenswrapper[5018]: I1014 08:53:28.625456 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd09ecc-4901-4232-b07b-b0e757534169" path="/var/lib/kubelet/pods/ccd09ecc-4901-4232-b07b-b0e757534169/volumes" Oct 14 08:53:29 crc kubenswrapper[5018]: I1014 08:53:29.069118 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:29 crc kubenswrapper[5018]: I1014 08:53:29.117575 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:53:29 crc kubenswrapper[5018]: W1014 08:53:29.124325 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05de60b4_2823_4420_b590_2e31ac05b0fe.slice/crio-cea5974d59471076cd0f7a6319db0c799042549cdb4bd3313bd3bf7b52ab9d78 WatchSource:0}: Error finding container cea5974d59471076cd0f7a6319db0c799042549cdb4bd3313bd3bf7b52ab9d78: Status 404 returned error can't find the container with id cea5974d59471076cd0f7a6319db0c799042549cdb4bd3313bd3bf7b52ab9d78 Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.094718 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerStarted","Data":"ef53fbe1eb381437837cbf0ca10ca7d3c3b0550ab75c3531ca2b8207894fbe27"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.095096 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerStarted","Data":"65b14e1c027a8cae67975b1fce4a4009ca6aec0752d8342f09ee93d4ac090487"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.095116 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerStarted","Data":"cea5974d59471076cd0f7a6319db0c799042549cdb4bd3313bd3bf7b52ab9d78"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.100557 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerStarted","Data":"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.100612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerStarted","Data":"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.100659 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerStarted","Data":"b1e2fbe493f934aa12474879dd7187a289060edd6fd48a043376f51d606daa14"} Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.143421 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.14339525 podStartE2EDuration="2.14339525s" podCreationTimestamp="2025-10-14 08:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:30.119765762 +0000 UTC m=+7426.703812449" watchObservedRunningTime="2025-10-14 08:53:30.14339525 +0000 UTC m=+7426.727441917" Oct 14 08:53:30 crc kubenswrapper[5018]: I1014 08:53:30.165496 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.165473234 podStartE2EDuration="2.165473234s" podCreationTimestamp="2025-10-14 08:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:30.163067085 +0000 UTC m=+7426.747113722" watchObservedRunningTime="2025-10-14 08:53:30.165473234 +0000 UTC m=+7426.749519871" Oct 14 08:53:32 crc kubenswrapper[5018]: E1014 08:53:32.048535 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:32 crc kubenswrapper[5018]: E1014 08:53:32.054322 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:32 crc kubenswrapper[5018]: E1014 08:53:32.056290 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:32 crc kubenswrapper[5018]: E1014 08:53:32.056387 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:33 crc kubenswrapper[5018]: I1014 08:53:33.558865 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:53:33 crc kubenswrapper[5018]: I1014 08:53:33.558916 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:53:37 crc kubenswrapper[5018]: E1014 08:53:37.049014 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:37 crc kubenswrapper[5018]: E1014 08:53:37.052807 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:37 crc kubenswrapper[5018]: E1014 08:53:37.055315 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:37 crc kubenswrapper[5018]: E1014 08:53:37.055413 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:38 crc kubenswrapper[5018]: I1014 08:53:38.551294 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:53:38 crc kubenswrapper[5018]: I1014 08:53:38.552044 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:53:38 crc kubenswrapper[5018]: I1014 08:53:38.558733 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:53:38 crc kubenswrapper[5018]: I1014 08:53:38.558813 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:53:39 crc kubenswrapper[5018]: I1014 08:53:39.645931 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.107:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:39 crc kubenswrapper[5018]: I1014 08:53:39.646467 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.108:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:39 crc kubenswrapper[5018]: I1014 08:53:39.646542 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.107:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:39 crc kubenswrapper[5018]: I1014 08:53:39.646599 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.108:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:53:42 crc kubenswrapper[5018]: E1014 08:53:42.047227 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:42 crc kubenswrapper[5018]: E1014 08:53:42.049008 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:42 crc kubenswrapper[5018]: E1014 08:53:42.050425 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:53:42 crc kubenswrapper[5018]: E1014 08:53:42.050486 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.276113 5018 generic.go:334] "Generic (PLEG): container finished" podID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" exitCode=137 Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.276120 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d87f8bc-88ae-442e-b342-1c888f199a53","Type":"ContainerDied","Data":"2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f"} Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.649817 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.737509 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ngqq\" (UniqueName: \"kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq\") pod \"8d87f8bc-88ae-442e-b342-1c888f199a53\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.737585 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data\") pod \"8d87f8bc-88ae-442e-b342-1c888f199a53\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.737705 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle\") pod \"8d87f8bc-88ae-442e-b342-1c888f199a53\" (UID: \"8d87f8bc-88ae-442e-b342-1c888f199a53\") " Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.744778 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq" (OuterVolumeSpecName: "kube-api-access-4ngqq") pod "8d87f8bc-88ae-442e-b342-1c888f199a53" (UID: "8d87f8bc-88ae-442e-b342-1c888f199a53"). InnerVolumeSpecName "kube-api-access-4ngqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.782452 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data" (OuterVolumeSpecName: "config-data") pod "8d87f8bc-88ae-442e-b342-1c888f199a53" (UID: "8d87f8bc-88ae-442e-b342-1c888f199a53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.787280 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d87f8bc-88ae-442e-b342-1c888f199a53" (UID: "8d87f8bc-88ae-442e-b342-1c888f199a53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.840558 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ngqq\" (UniqueName: \"kubernetes.io/projected/8d87f8bc-88ae-442e-b342-1c888f199a53-kube-api-access-4ngqq\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.841191 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:43 crc kubenswrapper[5018]: I1014 08:53:43.841343 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d87f8bc-88ae-442e-b342-1c888f199a53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.293222 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8d87f8bc-88ae-442e-b342-1c888f199a53","Type":"ContainerDied","Data":"34955393d349aee7ed2df928d1c500f78a222cf5176b1874939b75e5f4db92fe"} Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.293307 5018 scope.go:117] "RemoveContainer" containerID="2afc68fb161652c9d5a3aba12450b6c446d0fb4306dd51dae967bb3e5fd0545f" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.293312 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.353850 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.372580 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.421576 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:44 crc kubenswrapper[5018]: E1014 08:53:44.422101 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.422121 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.422294 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" containerName="nova-scheduler-scheduler" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.423280 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.428792 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.430269 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.558313 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll8h7\" (UniqueName: \"kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.558414 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.558677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.623909 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d87f8bc-88ae-442e-b342-1c888f199a53" path="/var/lib/kubelet/pods/8d87f8bc-88ae-442e-b342-1c888f199a53/volumes" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.660292 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.660601 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.660841 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll8h7\" (UniqueName: \"kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.663355 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.673941 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.678914 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.694303 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll8h7\" (UniqueName: \"kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7\") pod \"nova-scheduler-0\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " pod="openstack/nova-scheduler-0" Oct 14 08:53:44 crc kubenswrapper[5018]: I1014 08:53:44.745087 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:53:45 crc kubenswrapper[5018]: I1014 08:53:45.267516 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:53:45 crc kubenswrapper[5018]: I1014 08:53:45.312463 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8678903d-8a9e-466b-a43c-3aaed9c5e3a5","Type":"ContainerStarted","Data":"9af70c0d3318bdaa60014bb4acc2befa4ff36520ed41f356ac8dc928b223bb9f"} Oct 14 08:53:46 crc kubenswrapper[5018]: I1014 08:53:46.323923 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8678903d-8a9e-466b-a43c-3aaed9c5e3a5","Type":"ContainerStarted","Data":"bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789"} Oct 14 08:53:46 crc kubenswrapper[5018]: I1014 08:53:46.351818 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.351792034 podStartE2EDuration="2.351792034s" podCreationTimestamp="2025-10-14 08:53:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:46.349601911 +0000 UTC m=+7442.933648568" watchObservedRunningTime="2025-10-14 08:53:46.351792034 +0000 UTC m=+7442.935838701" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.553674 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.554554 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.555967 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.560541 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.573284 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.574958 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:53:48 crc kubenswrapper[5018]: I1014 08:53:48.587438 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.361233 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.366744 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.368344 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.581571 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.588577 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.609542 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.684976 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rwbw\" (UniqueName: \"kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.685036 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.685073 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.685107 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.685128 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.745417 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.786553 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rwbw\" (UniqueName: \"kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.786716 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.786823 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.786921 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.786999 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.787703 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.787749 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.787764 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.787874 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.803882 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rwbw\" (UniqueName: \"kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw\") pod \"dnsmasq-dns-655df9d575-9rwfg\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:49 crc kubenswrapper[5018]: I1014 08:53:49.911586 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:50 crc kubenswrapper[5018]: I1014 08:53:50.415759 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 08:53:50 crc kubenswrapper[5018]: W1014 08:53:50.422772 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55ca4b1_1c7a_49ea_bf68_c32a91de7c6e.slice/crio-0f06a849dc9a8e2dca36450a84975071fc78612e587deb6b7312e11a7085a9f2 WatchSource:0}: Error finding container 0f06a849dc9a8e2dca36450a84975071fc78612e587deb6b7312e11a7085a9f2: Status 404 returned error can't find the container with id 0f06a849dc9a8e2dca36450a84975071fc78612e587deb6b7312e11a7085a9f2 Oct 14 08:53:51 crc kubenswrapper[5018]: I1014 08:53:51.381024 5018 generic.go:334] "Generic (PLEG): container finished" podID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerID="1062635fc0651ed9abaa34283677781a36047d6343e5d6745ad4efbf76603dce" exitCode=0 Oct 14 08:53:51 crc kubenswrapper[5018]: I1014 08:53:51.382008 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" event={"ID":"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e","Type":"ContainerDied","Data":"1062635fc0651ed9abaa34283677781a36047d6343e5d6745ad4efbf76603dce"} Oct 14 08:53:51 crc kubenswrapper[5018]: I1014 08:53:51.382068 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" event={"ID":"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e","Type":"ContainerStarted","Data":"0f06a849dc9a8e2dca36450a84975071fc78612e587deb6b7312e11a7085a9f2"} Oct 14 08:53:52 crc kubenswrapper[5018]: I1014 08:53:52.245399 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:52 crc kubenswrapper[5018]: I1014 08:53:52.395055 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" event={"ID":"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e","Type":"ContainerStarted","Data":"5045a096ebb1e46ec923b34da32996ecc7663e4e86bfba2ce6bb474b4f341eec"} Oct 14 08:53:52 crc kubenswrapper[5018]: I1014 08:53:52.395299 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-log" containerID="cri-o://31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e" gracePeriod=30 Oct 14 08:53:52 crc kubenswrapper[5018]: I1014 08:53:52.395377 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-api" containerID="cri-o://68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c" gracePeriod=30 Oct 14 08:53:52 crc kubenswrapper[5018]: I1014 08:53:52.429535 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" podStartSLOduration=3.429515824 podStartE2EDuration="3.429515824s" podCreationTimestamp="2025-10-14 08:53:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:52.426661202 +0000 UTC m=+7449.010707839" watchObservedRunningTime="2025-10-14 08:53:52.429515824 +0000 UTC m=+7449.013562461" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.408980 5018 generic.go:334] "Generic (PLEG): container finished" podID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerID="31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e" exitCode=143 Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.409153 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerDied","Data":"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e"} Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.409390 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.665831 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.668335 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.706994 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.767817 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.767874 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.768014 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cz7g\" (UniqueName: \"kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.869806 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.869871 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.869984 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cz7g\" (UniqueName: \"kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.870830 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.870903 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:53 crc kubenswrapper[5018]: I1014 08:53:53.908392 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cz7g\" (UniqueName: \"kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g\") pod \"redhat-marketplace-v7jbb\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:54 crc kubenswrapper[5018]: I1014 08:53:54.026049 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:53:54 crc kubenswrapper[5018]: I1014 08:53:54.500370 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:53:54 crc kubenswrapper[5018]: W1014 08:53:54.501951 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda53e10b7_34b8_4576_80e8_a187ffe4b6a2.slice/crio-b6c2991a1782b06b2a72821a15c37f0aed43e719edb65a65434e288c84bf14bf WatchSource:0}: Error finding container b6c2991a1782b06b2a72821a15c37f0aed43e719edb65a65434e288c84bf14bf: Status 404 returned error can't find the container with id b6c2991a1782b06b2a72821a15c37f0aed43e719edb65a65434e288c84bf14bf Oct 14 08:53:54 crc kubenswrapper[5018]: I1014 08:53:54.745252 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 08:53:54 crc kubenswrapper[5018]: I1014 08:53:54.803517 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 08:53:55 crc kubenswrapper[5018]: I1014 08:53:55.440207 5018 generic.go:334] "Generic (PLEG): container finished" podID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerID="639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104" exitCode=0 Oct 14 08:53:55 crc kubenswrapper[5018]: I1014 08:53:55.440346 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerDied","Data":"639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104"} Oct 14 08:53:55 crc kubenswrapper[5018]: I1014 08:53:55.440712 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerStarted","Data":"b6c2991a1782b06b2a72821a15c37f0aed43e719edb65a65434e288c84bf14bf"} Oct 14 08:53:55 crc kubenswrapper[5018]: I1014 08:53:55.512303 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.171553 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.320158 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs\") pod \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.320503 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle\") pod \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.320545 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data\") pod \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.320582 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-262sj\" (UniqueName: \"kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj\") pod \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\" (UID: \"41205065-82c7-4c00-8faa-8dbc0da1c0aa\") " Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.321760 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs" (OuterVolumeSpecName: "logs") pod "41205065-82c7-4c00-8faa-8dbc0da1c0aa" (UID: "41205065-82c7-4c00-8faa-8dbc0da1c0aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.327631 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj" (OuterVolumeSpecName: "kube-api-access-262sj") pod "41205065-82c7-4c00-8faa-8dbc0da1c0aa" (UID: "41205065-82c7-4c00-8faa-8dbc0da1c0aa"). InnerVolumeSpecName "kube-api-access-262sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.366580 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data" (OuterVolumeSpecName: "config-data") pod "41205065-82c7-4c00-8faa-8dbc0da1c0aa" (UID: "41205065-82c7-4c00-8faa-8dbc0da1c0aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.370439 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41205065-82c7-4c00-8faa-8dbc0da1c0aa" (UID: "41205065-82c7-4c00-8faa-8dbc0da1c0aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.423062 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41205065-82c7-4c00-8faa-8dbc0da1c0aa-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.423089 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.423099 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41205065-82c7-4c00-8faa-8dbc0da1c0aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.423108 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-262sj\" (UniqueName: \"kubernetes.io/projected/41205065-82c7-4c00-8faa-8dbc0da1c0aa-kube-api-access-262sj\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.452263 5018 generic.go:334] "Generic (PLEG): container finished" podID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerID="c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a" exitCode=0 Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.452384 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerDied","Data":"c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a"} Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.457279 5018 generic.go:334] "Generic (PLEG): container finished" podID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerID="68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c" exitCode=0 Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.457736 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerDied","Data":"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c"} Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.457778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41205065-82c7-4c00-8faa-8dbc0da1c0aa","Type":"ContainerDied","Data":"b1e2fbe493f934aa12474879dd7187a289060edd6fd48a043376f51d606daa14"} Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.457779 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.457810 5018 scope.go:117] "RemoveContainer" containerID="68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.490715 5018 scope.go:117] "RemoveContainer" containerID="31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.497236 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.503824 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.517775 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:56 crc kubenswrapper[5018]: E1014 08:53:56.518512 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-api" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.518526 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-api" Oct 14 08:53:56 crc kubenswrapper[5018]: E1014 08:53:56.518561 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-log" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.518568 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-log" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.519711 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-api" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.519737 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" containerName="nova-api-log" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.522471 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.528558 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.530932 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.531261 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.531417 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.544819 5018 scope.go:117] "RemoveContainer" containerID="68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c" Oct 14 08:53:56 crc kubenswrapper[5018]: E1014 08:53:56.545446 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c\": container with ID starting with 68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c not found: ID does not exist" containerID="68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.545602 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c"} err="failed to get container status \"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c\": rpc error: code = NotFound desc = could not find container \"68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c\": container with ID starting with 68f4c2f48488abcb055c796f984f04d4df69a0bb6ee6885116c108bdf7fe116c not found: ID does not exist" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.545691 5018 scope.go:117] "RemoveContainer" containerID="31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e" Oct 14 08:53:56 crc kubenswrapper[5018]: E1014 08:53:56.546197 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e\": container with ID starting with 31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e not found: ID does not exist" containerID="31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.546447 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e"} err="failed to get container status \"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e\": rpc error: code = NotFound desc = could not find container \"31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e\": container with ID starting with 31280de6134c230b015c646147e301689ed3cdcea79727f3bf7080e4b6383f2e not found: ID does not exist" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.620197 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41205065-82c7-4c00-8faa-8dbc0da1c0aa" path="/var/lib/kubelet/pods/41205065-82c7-4c00-8faa-8dbc0da1c0aa/volumes" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642382 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642477 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642547 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642691 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642831 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.642970 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w87n2\" (UniqueName: \"kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.744611 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.744734 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.744794 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.744907 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w87n2\" (UniqueName: \"kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.745007 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.745080 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.746186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.749203 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.750991 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.752555 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.755584 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.764970 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w87n2\" (UniqueName: \"kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2\") pod \"nova-api-0\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " pod="openstack/nova-api-0" Oct 14 08:53:56 crc kubenswrapper[5018]: I1014 08:53:56.853244 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:53:57 crc kubenswrapper[5018]: I1014 08:53:57.406219 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:53:57 crc kubenswrapper[5018]: I1014 08:53:57.468351 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerStarted","Data":"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d"} Oct 14 08:53:57 crc kubenswrapper[5018]: I1014 08:53:57.474548 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerStarted","Data":"769a9ec21e4cbdd1dabaa7ffa6a86f902f0f3cdeb31c4045eedf45220331d074"} Oct 14 08:53:57 crc kubenswrapper[5018]: I1014 08:53:57.506264 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v7jbb" podStartSLOduration=2.974208089 podStartE2EDuration="4.506239199s" podCreationTimestamp="2025-10-14 08:53:53 +0000 UTC" firstStartedPulling="2025-10-14 08:53:55.444251966 +0000 UTC m=+7452.028298633" lastFinishedPulling="2025-10-14 08:53:56.976283116 +0000 UTC m=+7453.560329743" observedRunningTime="2025-10-14 08:53:57.488301024 +0000 UTC m=+7454.072347671" watchObservedRunningTime="2025-10-14 08:53:57.506239199 +0000 UTC m=+7454.090285836" Oct 14 08:53:58 crc kubenswrapper[5018]: I1014 08:53:58.491560 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerStarted","Data":"d4b9ecf9551b7d71f24978b3596fb93c674a49dbabf2d694c982d16be6b892f8"} Oct 14 08:53:58 crc kubenswrapper[5018]: I1014 08:53:58.492033 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerStarted","Data":"b5014c9bfac19f9acded35ea492bf1ac02255d3437039504d0647dddd1fdb140"} Oct 14 08:53:58 crc kubenswrapper[5018]: I1014 08:53:58.533260 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.533228131 podStartE2EDuration="2.533228131s" podCreationTimestamp="2025-10-14 08:53:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:53:58.521407281 +0000 UTC m=+7455.105453928" watchObservedRunningTime="2025-10-14 08:53:58.533228131 +0000 UTC m=+7455.117274798" Oct 14 08:53:59 crc kubenswrapper[5018]: I1014 08:53:59.913975 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.010516 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.010761 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="dnsmasq-dns" containerID="cri-o://dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c" gracePeriod=10 Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.504135 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.511754 5018 generic.go:334] "Generic (PLEG): container finished" podID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerID="dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c" exitCode=0 Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.511797 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" event={"ID":"9b01f3e1-d593-41e5-b338-4b6c07a24f2a","Type":"ContainerDied","Data":"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c"} Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.511805 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.511830 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74cc959f47-jbx4z" event={"ID":"9b01f3e1-d593-41e5-b338-4b6c07a24f2a","Type":"ContainerDied","Data":"33ac0f5bf92c5f6fde0fde6c62ad606fbb6db1597651847ee79ca755e0f680b7"} Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.511851 5018 scope.go:117] "RemoveContainer" containerID="dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.536317 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb\") pod \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.536556 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config\") pod \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.536650 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc\") pod \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.536747 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb\") pod \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.536785 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tttpn\" (UniqueName: \"kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn\") pod \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\" (UID: \"9b01f3e1-d593-41e5-b338-4b6c07a24f2a\") " Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.562784 5018 scope.go:117] "RemoveContainer" containerID="a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.576714 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn" (OuterVolumeSpecName: "kube-api-access-tttpn") pod "9b01f3e1-d593-41e5-b338-4b6c07a24f2a" (UID: "9b01f3e1-d593-41e5-b338-4b6c07a24f2a"). InnerVolumeSpecName "kube-api-access-tttpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.614300 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b01f3e1-d593-41e5-b338-4b6c07a24f2a" (UID: "9b01f3e1-d593-41e5-b338-4b6c07a24f2a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.618213 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b01f3e1-d593-41e5-b338-4b6c07a24f2a" (UID: "9b01f3e1-d593-41e5-b338-4b6c07a24f2a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.619419 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config" (OuterVolumeSpecName: "config") pod "9b01f3e1-d593-41e5-b338-4b6c07a24f2a" (UID: "9b01f3e1-d593-41e5-b338-4b6c07a24f2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.632980 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b01f3e1-d593-41e5-b338-4b6c07a24f2a" (UID: "9b01f3e1-d593-41e5-b338-4b6c07a24f2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.640645 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.640669 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.640678 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.640687 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.640710 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tttpn\" (UniqueName: \"kubernetes.io/projected/9b01f3e1-d593-41e5-b338-4b6c07a24f2a-kube-api-access-tttpn\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.687040 5018 scope.go:117] "RemoveContainer" containerID="dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c" Oct 14 08:54:00 crc kubenswrapper[5018]: E1014 08:54:00.687521 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c\": container with ID starting with dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c not found: ID does not exist" containerID="dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.687580 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c"} err="failed to get container status \"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c\": rpc error: code = NotFound desc = could not find container \"dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c\": container with ID starting with dccd37a74212cc16152c7639aaaa9b5469dc175b11bd9c43ab866b97daf9fc3c not found: ID does not exist" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.687609 5018 scope.go:117] "RemoveContainer" containerID="a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f" Oct 14 08:54:00 crc kubenswrapper[5018]: E1014 08:54:00.688075 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f\": container with ID starting with a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f not found: ID does not exist" containerID="a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.688108 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f"} err="failed to get container status \"a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f\": rpc error: code = NotFound desc = could not find container \"a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f\": container with ID starting with a03825479289e9168d4667249789e862918ba755d78c52ee9eb2a22d8636df5f not found: ID does not exist" Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.847498 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:54:00 crc kubenswrapper[5018]: I1014 08:54:00.854686 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74cc959f47-jbx4z"] Oct 14 08:54:02 crc kubenswrapper[5018]: I1014 08:54:02.625070 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" path="/var/lib/kubelet/pods/9b01f3e1-d593-41e5-b338-4b6c07a24f2a/volumes" Oct 14 08:54:04 crc kubenswrapper[5018]: I1014 08:54:04.026147 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:04 crc kubenswrapper[5018]: I1014 08:54:04.026501 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:04 crc kubenswrapper[5018]: I1014 08:54:04.116211 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:04 crc kubenswrapper[5018]: I1014 08:54:04.639733 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:04 crc kubenswrapper[5018]: I1014 08:54:04.728780 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:54:06 crc kubenswrapper[5018]: I1014 08:54:06.588519 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v7jbb" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="registry-server" containerID="cri-o://5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d" gracePeriod=2 Oct 14 08:54:06 crc kubenswrapper[5018]: E1014 08:54:06.668656 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda53e10b7_34b8_4576_80e8_a187ffe4b6a2.slice/crio-conmon-5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:54:06 crc kubenswrapper[5018]: I1014 08:54:06.853926 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:54:06 crc kubenswrapper[5018]: I1014 08:54:06.853974 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.048760 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.095040 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities\") pod \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.095115 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content\") pod \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.095205 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cz7g\" (UniqueName: \"kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g\") pod \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\" (UID: \"a53e10b7-34b8-4576-80e8-a187ffe4b6a2\") " Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.095798 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities" (OuterVolumeSpecName: "utilities") pod "a53e10b7-34b8-4576-80e8-a187ffe4b6a2" (UID: "a53e10b7-34b8-4576-80e8-a187ffe4b6a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.096440 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.101979 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g" (OuterVolumeSpecName: "kube-api-access-8cz7g") pod "a53e10b7-34b8-4576-80e8-a187ffe4b6a2" (UID: "a53e10b7-34b8-4576-80e8-a187ffe4b6a2"). InnerVolumeSpecName "kube-api-access-8cz7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.109506 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a53e10b7-34b8-4576-80e8-a187ffe4b6a2" (UID: "a53e10b7-34b8-4576-80e8-a187ffe4b6a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.197751 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.197787 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cz7g\" (UniqueName: \"kubernetes.io/projected/a53e10b7-34b8-4576-80e8-a187ffe4b6a2-kube-api-access-8cz7g\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.603474 5018 generic.go:334] "Generic (PLEG): container finished" podID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerID="5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d" exitCode=0 Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.603612 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerDied","Data":"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d"} Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.603990 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v7jbb" event={"ID":"a53e10b7-34b8-4576-80e8-a187ffe4b6a2","Type":"ContainerDied","Data":"b6c2991a1782b06b2a72821a15c37f0aed43e719edb65a65434e288c84bf14bf"} Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.604668 5018 scope.go:117] "RemoveContainer" containerID="5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.603704 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v7jbb" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.657065 5018 scope.go:117] "RemoveContainer" containerID="c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.663246 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.675040 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v7jbb"] Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.697977 5018 scope.go:117] "RemoveContainer" containerID="639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.755699 5018 scope.go:117] "RemoveContainer" containerID="5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d" Oct 14 08:54:07 crc kubenswrapper[5018]: E1014 08:54:07.756180 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d\": container with ID starting with 5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d not found: ID does not exist" containerID="5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.756224 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d"} err="failed to get container status \"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d\": rpc error: code = NotFound desc = could not find container \"5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d\": container with ID starting with 5e370631c27e574311ba80f3d70ecbf9422a01cf1dfdce46e102fc897ef7d56d not found: ID does not exist" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.756250 5018 scope.go:117] "RemoveContainer" containerID="c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a" Oct 14 08:54:07 crc kubenswrapper[5018]: E1014 08:54:07.756607 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a\": container with ID starting with c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a not found: ID does not exist" containerID="c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.756641 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a"} err="failed to get container status \"c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a\": rpc error: code = NotFound desc = could not find container \"c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a\": container with ID starting with c03484ae5e9cc39e0581fba019b0df660ce9fd49ff7422b264ca1a96a5a85a9a not found: ID does not exist" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.756655 5018 scope.go:117] "RemoveContainer" containerID="639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104" Oct 14 08:54:07 crc kubenswrapper[5018]: E1014 08:54:07.757012 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104\": container with ID starting with 639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104 not found: ID does not exist" containerID="639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.757034 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104"} err="failed to get container status \"639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104\": rpc error: code = NotFound desc = could not find container \"639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104\": container with ID starting with 639e7be054defdef1ea248c0f863152b4aef69d843595eac7a0060f40061a104 not found: ID does not exist" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.869901 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.112:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:54:07 crc kubenswrapper[5018]: I1014 08:54:07.869917 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.112:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:54:08 crc kubenswrapper[5018]: I1014 08:54:08.628750 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" path="/var/lib/kubelet/pods/a53e10b7-34b8-4576-80e8-a187ffe4b6a2/volumes" Oct 14 08:54:16 crc kubenswrapper[5018]: I1014 08:54:16.859181 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:54:16 crc kubenswrapper[5018]: I1014 08:54:16.860495 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:54:16 crc kubenswrapper[5018]: I1014 08:54:16.862726 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:54:16 crc kubenswrapper[5018]: I1014 08:54:16.871459 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:54:17 crc kubenswrapper[5018]: I1014 08:54:17.733883 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:54:17 crc kubenswrapper[5018]: I1014 08:54:17.747181 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.008860 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:54:30 crc kubenswrapper[5018]: E1014 08:54:30.011930 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="dnsmasq-dns" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012072 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="dnsmasq-dns" Oct 14 08:54:30 crc kubenswrapper[5018]: E1014 08:54:30.012100 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="extract-content" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012106 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="extract-content" Oct 14 08:54:30 crc kubenswrapper[5018]: E1014 08:54:30.012126 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="registry-server" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012134 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="registry-server" Oct 14 08:54:30 crc kubenswrapper[5018]: E1014 08:54:30.012144 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="extract-utilities" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012150 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="extract-utilities" Oct 14 08:54:30 crc kubenswrapper[5018]: E1014 08:54:30.012167 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="init" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012172 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="init" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012324 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53e10b7-34b8-4576-80e8-a187ffe4b6a2" containerName="registry-server" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.012346 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b01f3e1-d593-41e5-b338-4b6c07a24f2a" containerName="dnsmasq-dns" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.013335 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.017344 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-sw76v" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.017757 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.019863 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.020092 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.022268 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.066931 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.067150 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-log" containerID="cri-o://1a6faf8f27efe70c6e8bc2976dade19360f33221cf52aa54557011791f8c78a8" gracePeriod=30 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.067548 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-httpd" containerID="cri-o://c033e84e1d5c677e33aa151740e1f8f7eda2848c801a1b235d43faa5289af0f2" gracePeriod=30 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.115713 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.115784 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.115840 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.116018 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdpfr\" (UniqueName: \"kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.116067 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.164156 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.165958 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.198703 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.198944 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-log" containerID="cri-o://41cc3322c56458de6f5bcf72edcc9e797ada50be777ccec9a283724dbd865ee8" gracePeriod=30 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.199021 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-httpd" containerID="cri-o://295a83f13373ddcea2ff47866b73fdb6d1f09e67500d56c6ab48553e5df24d71" gracePeriod=30 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.212558 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.217611 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdpfr\" (UniqueName: \"kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.217708 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.217771 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.217820 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.217881 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.220993 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.221852 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.222592 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.234347 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.239853 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdpfr\" (UniqueName: \"kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr\") pod \"horizon-789f4f677c-vvw6r\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.319546 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28msk\" (UniqueName: \"kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.319935 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.319993 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.320167 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.320340 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.340642 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.421976 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422059 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422124 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422194 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28msk\" (UniqueName: \"kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422269 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422638 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.422897 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.424183 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.429084 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.453079 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28msk\" (UniqueName: \"kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk\") pod \"horizon-67ff97b49c-79fh8\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.650982 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.817593 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.903879 5018 generic.go:334] "Generic (PLEG): container finished" podID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerID="41cc3322c56458de6f5bcf72edcc9e797ada50be777ccec9a283724dbd865ee8" exitCode=143 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.903948 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerDied","Data":"41cc3322c56458de6f5bcf72edcc9e797ada50be777ccec9a283724dbd865ee8"} Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.904951 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerStarted","Data":"869aafa8b604261256e436db7b6ff1db2f4968e8a4b30150a818ed6cb57103b7"} Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.906850 5018 generic.go:334] "Generic (PLEG): container finished" podID="50595d02-681b-4587-83ce-66e5d5c69951" containerID="1a6faf8f27efe70c6e8bc2976dade19360f33221cf52aa54557011791f8c78a8" exitCode=143 Oct 14 08:54:30 crc kubenswrapper[5018]: I1014 08:54:30.906874 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerDied","Data":"1a6faf8f27efe70c6e8bc2976dade19360f33221cf52aa54557011791f8c78a8"} Oct 14 08:54:31 crc kubenswrapper[5018]: I1014 08:54:31.209935 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:54:31 crc kubenswrapper[5018]: W1014 08:54:31.216066 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266 WatchSource:0}: Error finding container 944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266: Status 404 returned error can't find the container with id 944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266 Oct 14 08:54:31 crc kubenswrapper[5018]: I1014 08:54:31.930809 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerStarted","Data":"944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266"} Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.113717 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.162785 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.165608 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.173662 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174210 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174274 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174301 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174578 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174780 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174800 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.174856 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-879tj\" (UniqueName: \"kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.183198 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.194795 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.229051 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.232725 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.257024 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.276843 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.276924 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.276990 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277012 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277027 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277085 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-879tj\" (UniqueName: \"kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277127 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277157 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277242 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277298 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277328 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277351 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.277422 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dx4z\" (UniqueName: \"kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.278094 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.278545 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.279675 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.282980 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.283096 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.283677 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.295158 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-879tj\" (UniqueName: \"kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj\") pod \"horizon-99549d576-lsj55\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.378970 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379044 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379096 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379122 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379142 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379178 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dx4z\" (UniqueName: \"kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379225 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379413 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.379954 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.381077 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.382693 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.382955 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.383751 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.393998 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dx4z\" (UniqueName: \"kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z\") pod \"horizon-5cdc789c6-8thlf\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.502824 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.560544 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:32 crc kubenswrapper[5018]: I1014 08:54:32.938339 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:54:32 crc kubenswrapper[5018]: W1014 08:54:32.940186 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f090ce_afb6_4d48_bf6e_3ad593315c84.slice/crio-e0f32af6a386824c1dde631598991f5cde7299c0b834e59dc61863f74975748d WatchSource:0}: Error finding container e0f32af6a386824c1dde631598991f5cde7299c0b834e59dc61863f74975748d: Status 404 returned error can't find the container with id e0f32af6a386824c1dde631598991f5cde7299c0b834e59dc61863f74975748d Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.054466 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:54:33 crc kubenswrapper[5018]: W1014 08:54:33.061190 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e10a610_e422_4785_89a4_5c1beb50c116.slice/crio-05313d0a7ff759c703224c1e949fb44c7979dfebe4f6685fd7450d5381fcd431 WatchSource:0}: Error finding container 05313d0a7ff759c703224c1e949fb44c7979dfebe4f6685fd7450d5381fcd431: Status 404 returned error can't find the container with id 05313d0a7ff759c703224c1e949fb44c7979dfebe4f6685fd7450d5381fcd431 Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.947393 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerStarted","Data":"05313d0a7ff759c703224c1e949fb44c7979dfebe4f6685fd7450d5381fcd431"} Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.948299 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerStarted","Data":"e0f32af6a386824c1dde631598991f5cde7299c0b834e59dc61863f74975748d"} Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.951241 5018 generic.go:334] "Generic (PLEG): container finished" podID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerID="295a83f13373ddcea2ff47866b73fdb6d1f09e67500d56c6ab48553e5df24d71" exitCode=0 Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.951320 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerDied","Data":"295a83f13373ddcea2ff47866b73fdb6d1f09e67500d56c6ab48553e5df24d71"} Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.951385 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e","Type":"ContainerDied","Data":"ca3f3cbc9456e3b957114c2cff5dd46d4b2ff7159e09a6329939bab86aa78bad"} Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.951396 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca3f3cbc9456e3b957114c2cff5dd46d4b2ff7159e09a6329939bab86aa78bad" Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.953161 5018 generic.go:334] "Generic (PLEG): container finished" podID="50595d02-681b-4587-83ce-66e5d5c69951" containerID="c033e84e1d5c677e33aa151740e1f8f7eda2848c801a1b235d43faa5289af0f2" exitCode=0 Oct 14 08:54:33 crc kubenswrapper[5018]: I1014 08:54:33.953187 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerDied","Data":"c033e84e1d5c677e33aa151740e1f8f7eda2848c801a1b235d43faa5289af0f2"} Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:33.998209 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.020970 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqj9w\" (UniqueName: \"kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.021255 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.021330 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.021362 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.022171 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs" (OuterVolumeSpecName: "logs") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.022342 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.022675 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.022748 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.023254 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs\") pod \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\" (UID: \"1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.024214 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.024233 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.031806 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts" (OuterVolumeSpecName: "scripts") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.037666 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w" (OuterVolumeSpecName: "kube-api-access-sqj9w") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "kube-api-access-sqj9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.081636 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.095324 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.099676 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data" (OuterVolumeSpecName: "config-data") pod "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" (UID: "1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.102790 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.125543 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.126005 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.126082 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.126219 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45754\" (UniqueName: \"kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.126324 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.127769 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.127865 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs\") pod \"50595d02-681b-4587-83ce-66e5d5c69951\" (UID: \"50595d02-681b-4587-83ce-66e5d5c69951\") " Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.128320 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.128379 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.128440 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.128494 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.128546 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqj9w\" (UniqueName: \"kubernetes.io/projected/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e-kube-api-access-sqj9w\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.130388 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs" (OuterVolumeSpecName: "logs") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.131843 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.136579 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts" (OuterVolumeSpecName: "scripts") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.138275 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754" (OuterVolumeSpecName: "kube-api-access-45754") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "kube-api-access-45754". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.165834 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.182064 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.189292 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data" (OuterVolumeSpecName: "config-data") pod "50595d02-681b-4587-83ce-66e5d5c69951" (UID: "50595d02-681b-4587-83ce-66e5d5c69951"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230331 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45754\" (UniqueName: \"kubernetes.io/projected/50595d02-681b-4587-83ce-66e5d5c69951-kube-api-access-45754\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230360 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230417 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230439 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230449 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50595d02-681b-4587-83ce-66e5d5c69951-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230457 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.230467 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50595d02-681b-4587-83ce-66e5d5c69951-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.969153 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.969311 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.969327 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50595d02-681b-4587-83ce-66e5d5c69951","Type":"ContainerDied","Data":"79c024816c81de2ef7a8fd22e80c3a34a1f041cd3f1c0e336e6e1b81c5488052"} Oct 14 08:54:34 crc kubenswrapper[5018]: I1014 08:54:34.969585 5018 scope.go:117] "RemoveContainer" containerID="c033e84e1d5c677e33aa151740e1f8f7eda2848c801a1b235d43faa5289af0f2" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.020047 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.031774 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.058789 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.076577 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.085629 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: E1014 08:54:35.086411 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086434 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: E1014 08:54:35.086459 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086466 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: E1014 08:54:35.086474 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086482 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: E1014 08:54:35.086505 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086514 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086897 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086949 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-log" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086963 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="50595d02-681b-4587-83ce-66e5d5c69951" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.086971 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" containerName="glance-httpd" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.093250 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.095376 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.098122 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.098308 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.098413 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zdjgk" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.099019 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.103070 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.105234 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.108936 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.110673 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.114781 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150396 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2kg9\" (UniqueName: \"kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150437 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150471 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150538 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgt26\" (UniqueName: \"kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150555 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150570 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150667 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150713 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150738 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150847 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.150984 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.151029 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.151044 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.151131 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.252981 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253050 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253069 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253474 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253538 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2kg9\" (UniqueName: \"kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253556 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253586 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253635 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgt26\" (UniqueName: \"kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253830 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253851 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253866 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253723 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253884 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.253989 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.254146 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.254869 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.254866 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.257545 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.257915 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.259525 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.259843 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.261976 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.264333 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.271242 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.271578 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2kg9\" (UniqueName: \"kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.273030 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgt26\" (UniqueName: \"kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26\") pod \"glance-default-internal-api-0\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.281587 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.281955 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.425105 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:54:35 crc kubenswrapper[5018]: I1014 08:54:35.439951 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:36 crc kubenswrapper[5018]: I1014 08:54:36.650663 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e" path="/var/lib/kubelet/pods/1e88c1a2-99a0-4d22-81a5-fbd7c6a5d06e/volumes" Oct 14 08:54:36 crc kubenswrapper[5018]: I1014 08:54:36.656057 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50595d02-681b-4587-83ce-66e5d5c69951" path="/var/lib/kubelet/pods/50595d02-681b-4587-83ce-66e5d5c69951/volumes" Oct 14 08:54:39 crc kubenswrapper[5018]: I1014 08:54:39.404193 5018 scope.go:117] "RemoveContainer" containerID="1a6faf8f27efe70c6e8bc2976dade19360f33221cf52aa54557011791f8c78a8" Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.000631 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.050168 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerStarted","Data":"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1"} Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.052059 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerStarted","Data":"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb"} Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.054480 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerStarted","Data":"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c"} Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.057755 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerStarted","Data":"0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54"} Oct 14 08:54:40 crc kubenswrapper[5018]: W1014 08:54:40.072035 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc347af0_75b9_4175_bc5e_70f3078e35c6.slice/crio-6aa0fb196e8892875265f3b590d03151715a0a279f51639d7b81ab50ccf134c4 WatchSource:0}: Error finding container 6aa0fb196e8892875265f3b590d03151715a0a279f51639d7b81ab50ccf134c4: Status 404 returned error can't find the container with id 6aa0fb196e8892875265f3b590d03151715a0a279f51639d7b81ab50ccf134c4 Oct 14 08:54:40 crc kubenswrapper[5018]: I1014 08:54:40.146572 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.076941 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerStarted","Data":"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.079419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerStarted","Data":"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.083724 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerStarted","Data":"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.083831 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67ff97b49c-79fh8" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon-log" containerID="cri-o://ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" gracePeriod=30 Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.084028 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67ff97b49c-79fh8" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon" containerID="cri-o://0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" gracePeriod=30 Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.086219 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerStarted","Data":"b10bff320cd1aef3f540117705e3e341d22490e36c13290cfba7eec50e8f6a71"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.086241 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerStarted","Data":"38a87e8b8e86e3dd91c7651f3591d41a64c55b22d38f22a130a6f83b2a0b8979"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.087526 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerStarted","Data":"ba05b131d9cc289cc0f7d05bb2b92055b3494f32ba43fa2269cb65376bf6bbb9"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.087549 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerStarted","Data":"6aa0fb196e8892875265f3b590d03151715a0a279f51639d7b81ab50ccf134c4"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.099761 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerStarted","Data":"ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5"} Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.099904 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-789f4f677c-vvw6r" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon-log" containerID="cri-o://0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54" gracePeriod=30 Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.100135 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-789f4f677c-vvw6r" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon" containerID="cri-o://ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5" gracePeriod=30 Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.110096 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5cdc789c6-8thlf" podStartSLOduration=2.484588389 podStartE2EDuration="9.110077413s" podCreationTimestamp="2025-10-14 08:54:32 +0000 UTC" firstStartedPulling="2025-10-14 08:54:33.063678199 +0000 UTC m=+7489.647724826" lastFinishedPulling="2025-10-14 08:54:39.689167223 +0000 UTC m=+7496.273213850" observedRunningTime="2025-10-14 08:54:41.096976157 +0000 UTC m=+7497.681022784" watchObservedRunningTime="2025-10-14 08:54:41.110077413 +0000 UTC m=+7497.694124030" Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.116359 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-99549d576-lsj55" podStartSLOduration=2.439296319 podStartE2EDuration="9.116343663s" podCreationTimestamp="2025-10-14 08:54:32 +0000 UTC" firstStartedPulling="2025-10-14 08:54:32.943325864 +0000 UTC m=+7489.527372491" lastFinishedPulling="2025-10-14 08:54:39.620373208 +0000 UTC m=+7496.204419835" observedRunningTime="2025-10-14 08:54:41.114971303 +0000 UTC m=+7497.699017940" watchObservedRunningTime="2025-10-14 08:54:41.116343663 +0000 UTC m=+7497.700390300" Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.151593 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67ff97b49c-79fh8" podStartSLOduration=2.7515662990000003 podStartE2EDuration="11.151570964s" podCreationTimestamp="2025-10-14 08:54:30 +0000 UTC" firstStartedPulling="2025-10-14 08:54:31.21921547 +0000 UTC m=+7487.803262097" lastFinishedPulling="2025-10-14 08:54:39.619220135 +0000 UTC m=+7496.203266762" observedRunningTime="2025-10-14 08:54:41.145457208 +0000 UTC m=+7497.729503845" watchObservedRunningTime="2025-10-14 08:54:41.151570964 +0000 UTC m=+7497.735617591" Oct 14 08:54:41 crc kubenswrapper[5018]: I1014 08:54:41.171178 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-789f4f677c-vvw6r" podStartSLOduration=3.392075549 podStartE2EDuration="12.171159076s" podCreationTimestamp="2025-10-14 08:54:29 +0000 UTC" firstStartedPulling="2025-10-14 08:54:30.838132141 +0000 UTC m=+7487.422178768" lastFinishedPulling="2025-10-14 08:54:39.617215658 +0000 UTC m=+7496.201262295" observedRunningTime="2025-10-14 08:54:41.166605156 +0000 UTC m=+7497.750651783" watchObservedRunningTime="2025-10-14 08:54:41.171159076 +0000 UTC m=+7497.755205703" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.111449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerStarted","Data":"89c2c3c1c29bf6244a053cff8a3b895dd672b6f514f9287e56db1eeba83bd95e"} Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.116094 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerStarted","Data":"d43f95354a677e6c3aafe9f165dbe59e283b3bddc5f2df6c04584560150ca6b8"} Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.135801 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.135784567 podStartE2EDuration="7.135784567s" podCreationTimestamp="2025-10-14 08:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:54:42.129207858 +0000 UTC m=+7498.713254485" watchObservedRunningTime="2025-10-14 08:54:42.135784567 +0000 UTC m=+7498.719831194" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.151131 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.151117027 podStartE2EDuration="8.151117027s" podCreationTimestamp="2025-10-14 08:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:54:42.149546512 +0000 UTC m=+7498.733593139" watchObservedRunningTime="2025-10-14 08:54:42.151117027 +0000 UTC m=+7498.735163654" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.503863 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.503962 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.561352 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:42 crc kubenswrapper[5018]: I1014 08:54:42.561411 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.426335 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.426931 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.440539 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.440565 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.466525 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.477822 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.487156 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:45 crc kubenswrapper[5018]: I1014 08:54:45.492143 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:46 crc kubenswrapper[5018]: I1014 08:54:46.166287 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:46 crc kubenswrapper[5018]: I1014 08:54:46.166513 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:54:46 crc kubenswrapper[5018]: I1014 08:54:46.166665 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:46 crc kubenswrapper[5018]: I1014 08:54:46.166689 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.186153 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.186525 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.186875 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.189053 5018 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.227932 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.389562 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.398823 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:54:48 crc kubenswrapper[5018]: I1014 08:54:48.622227 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:54:50 crc kubenswrapper[5018]: I1014 08:54:50.056135 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-htt58"] Oct 14 08:54:50 crc kubenswrapper[5018]: I1014 08:54:50.073086 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-htt58"] Oct 14 08:54:50 crc kubenswrapper[5018]: I1014 08:54:50.341342 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:54:50 crc kubenswrapper[5018]: I1014 08:54:50.620253 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf2f6095-4fe6-43fa-a482-5767e720f595" path="/var/lib/kubelet/pods/cf2f6095-4fe6-43fa-a482-5767e720f595/volumes" Oct 14 08:54:50 crc kubenswrapper[5018]: I1014 08:54:50.656123 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:54:52 crc kubenswrapper[5018]: I1014 08:54:52.508183 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.115:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8443: connect: connection refused" Oct 14 08:54:52 crc kubenswrapper[5018]: I1014 08:54:52.562971 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.116:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8443: connect: connection refused" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.242169 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.245173 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.256768 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.368514 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.368709 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.368771 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr75h\" (UniqueName: \"kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.471086 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.471242 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.471308 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr75h\" (UniqueName: \"kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.471754 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.471848 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.490427 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr75h\" (UniqueName: \"kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h\") pod \"community-operators-g4jf6\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:57 crc kubenswrapper[5018]: I1014 08:54:57.583185 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:54:58 crc kubenswrapper[5018]: I1014 08:54:58.119173 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:54:58 crc kubenswrapper[5018]: I1014 08:54:58.293242 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerStarted","Data":"d08aaa102ef49015f622f49d0c6f2b9b63a878402b2e757e62c157cb81abf7df"} Oct 14 08:54:59 crc kubenswrapper[5018]: I1014 08:54:59.302796 5018 generic.go:334] "Generic (PLEG): container finished" podID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerID="8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595" exitCode=0 Oct 14 08:54:59 crc kubenswrapper[5018]: I1014 08:54:59.302979 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerDied","Data":"8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595"} Oct 14 08:55:00 crc kubenswrapper[5018]: I1014 08:55:00.045186 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3632-account-create-94rwz"] Oct 14 08:55:00 crc kubenswrapper[5018]: I1014 08:55:00.069424 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3632-account-create-94rwz"] Oct 14 08:55:00 crc kubenswrapper[5018]: I1014 08:55:00.626476 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9a074e-cd7c-4cb9-bde2-fe324ab51708" path="/var/lib/kubelet/pods/ee9a074e-cd7c-4cb9-bde2-fe324ab51708/volumes" Oct 14 08:55:01 crc kubenswrapper[5018]: I1014 08:55:01.331332 5018 generic.go:334] "Generic (PLEG): container finished" podID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerID="4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc" exitCode=0 Oct 14 08:55:01 crc kubenswrapper[5018]: I1014 08:55:01.331450 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerDied","Data":"4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc"} Oct 14 08:55:02 crc kubenswrapper[5018]: I1014 08:55:02.463401 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:55:02 crc kubenswrapper[5018]: I1014 08:55:02.463852 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:55:03 crc kubenswrapper[5018]: I1014 08:55:03.368432 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerStarted","Data":"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c"} Oct 14 08:55:03 crc kubenswrapper[5018]: I1014 08:55:03.395202 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g4jf6" podStartSLOduration=3.349508078 podStartE2EDuration="6.395173957s" podCreationTimestamp="2025-10-14 08:54:57 +0000 UTC" firstStartedPulling="2025-10-14 08:54:59.304603008 +0000 UTC m=+7515.888649635" lastFinishedPulling="2025-10-14 08:55:02.350268857 +0000 UTC m=+7518.934315514" observedRunningTime="2025-10-14 08:55:03.389008832 +0000 UTC m=+7519.973055459" watchObservedRunningTime="2025-10-14 08:55:03.395173957 +0000 UTC m=+7519.979220594" Oct 14 08:55:04 crc kubenswrapper[5018]: I1014 08:55:04.685359 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:55:04 crc kubenswrapper[5018]: I1014 08:55:04.902105 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:55:06 crc kubenswrapper[5018]: I1014 08:55:06.474194 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:55:06 crc kubenswrapper[5018]: I1014 08:55:06.631360 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:55:06 crc kubenswrapper[5018]: I1014 08:55:06.695971 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:55:07 crc kubenswrapper[5018]: I1014 08:55:07.413214 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon-log" containerID="cri-o://7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb" gracePeriod=30 Oct 14 08:55:07 crc kubenswrapper[5018]: I1014 08:55:07.413319 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" containerID="cri-o://7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc" gracePeriod=30 Oct 14 08:55:07 crc kubenswrapper[5018]: I1014 08:55:07.583875 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:07 crc kubenswrapper[5018]: I1014 08:55:07.583960 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:07 crc kubenswrapper[5018]: I1014 08:55:07.663270 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:08 crc kubenswrapper[5018]: I1014 08:55:08.494255 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:08 crc kubenswrapper[5018]: I1014 08:55:08.571554 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:55:09 crc kubenswrapper[5018]: I1014 08:55:09.547113 5018 scope.go:117] "RemoveContainer" containerID="79d84bb2589c90bf3af60e2e4c20a9e0c372807e8dd8a0afc53500cc2f00f05b" Oct 14 08:55:09 crc kubenswrapper[5018]: I1014 08:55:09.585949 5018 scope.go:117] "RemoveContainer" containerID="ed974f0484876dd1fa4b142f26956a40a8b3537cf9521a9601f07815d7625c75" Oct 14 08:55:10 crc kubenswrapper[5018]: I1014 08:55:10.444721 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g4jf6" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="registry-server" containerID="cri-o://7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c" gracePeriod=2 Oct 14 08:55:10 crc kubenswrapper[5018]: I1014 08:55:10.965557 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.109729 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities\") pod \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.109799 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content\") pod \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.109843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr75h\" (UniqueName: \"kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h\") pod \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\" (UID: \"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.110501 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities" (OuterVolumeSpecName: "utilities") pod "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" (UID: "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.118244 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h" (OuterVolumeSpecName: "kube-api-access-gr75h") pod "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" (UID: "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f"). InnerVolumeSpecName "kube-api-access-gr75h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.173667 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" (UID: "dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.212144 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.212176 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr75h\" (UniqueName: \"kubernetes.io/projected/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-kube-api-access-gr75h\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.212190 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.404379 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.464236 5018 generic.go:334] "Generic (PLEG): container finished" podID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerID="7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c" exitCode=0 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.464323 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerDied","Data":"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.464351 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jf6" event={"ID":"dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f","Type":"ContainerDied","Data":"d08aaa102ef49015f622f49d0c6f2b9b63a878402b2e757e62c157cb81abf7df"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.464363 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jf6" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.464368 5018 scope.go:117] "RemoveContainer" containerID="7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.468769 5018 generic.go:334] "Generic (PLEG): container finished" podID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerID="7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc" exitCode=0 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.468833 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerDied","Data":"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471867 5018 generic.go:334] "Generic (PLEG): container finished" podID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerID="0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" exitCode=137 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471890 5018 generic.go:334] "Generic (PLEG): container finished" podID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerID="ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" exitCode=137 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471915 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67ff97b49c-79fh8" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471936 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerDied","Data":"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471968 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerDied","Data":"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.471984 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67ff97b49c-79fh8" event={"ID":"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974","Type":"ContainerDied","Data":"944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.473758 5018 generic.go:334] "Generic (PLEG): container finished" podID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerID="ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5" exitCode=137 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.473780 5018 generic.go:334] "Generic (PLEG): container finished" podID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerID="0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54" exitCode=137 Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.473794 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerDied","Data":"ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.473808 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerDied","Data":"0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54"} Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.487052 5018 scope.go:117] "RemoveContainer" containerID="4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.504997 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.506099 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.516201 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g4jf6"] Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.518521 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data\") pod \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.518565 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs\") pod \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.518670 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28msk\" (UniqueName: \"kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk\") pod \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.518734 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key\") pod \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.518796 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts\") pod \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\" (UID: \"9b5eed1f-c3f5-4a14-bf72-aaa3338dd974\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.519613 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs" (OuterVolumeSpecName: "logs") pod "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" (UID: "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.526025 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" (UID: "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.533304 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk" (OuterVolumeSpecName: "kube-api-access-28msk") pod "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" (UID: "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974"). InnerVolumeSpecName "kube-api-access-28msk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.534073 5018 scope.go:117] "RemoveContainer" containerID="8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.560706 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts" (OuterVolumeSpecName: "scripts") pod "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" (UID: "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.567191 5018 scope.go:117] "RemoveContainer" containerID="7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c" Oct 14 08:55:11 crc kubenswrapper[5018]: E1014 08:55:11.567924 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c\": container with ID starting with 7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c not found: ID does not exist" containerID="7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.567977 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c"} err="failed to get container status \"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c\": rpc error: code = NotFound desc = could not find container \"7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c\": container with ID starting with 7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.568007 5018 scope.go:117] "RemoveContainer" containerID="4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc" Oct 14 08:55:11 crc kubenswrapper[5018]: E1014 08:55:11.568747 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc\": container with ID starting with 4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc not found: ID does not exist" containerID="4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.568797 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc"} err="failed to get container status \"4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc\": rpc error: code = NotFound desc = could not find container \"4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc\": container with ID starting with 4f3b29482937515cdf120e691e022343b6618365837979fdb2d88f34b3cbccdc not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.568829 5018 scope.go:117] "RemoveContainer" containerID="8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595" Oct 14 08:55:11 crc kubenswrapper[5018]: E1014 08:55:11.569702 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595\": container with ID starting with 8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595 not found: ID does not exist" containerID="8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.569763 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595"} err="failed to get container status \"8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595\": rpc error: code = NotFound desc = could not find container \"8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595\": container with ID starting with 8bd7cd08c9446277f27cce47ad73c89ec542e7c5de862bf3535eae4278506595 not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.569788 5018 scope.go:117] "RemoveContainer" containerID="0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.570562 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data" (OuterVolumeSpecName: "config-data") pod "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" (UID: "9b5eed1f-c3f5-4a14-bf72-aaa3338dd974"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.620824 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts\") pod \"20e59f94-8dcb-4647-ad28-3f9b082392cb\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.620888 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs\") pod \"20e59f94-8dcb-4647-ad28-3f9b082392cb\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.621216 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs" (OuterVolumeSpecName: "logs") pod "20e59f94-8dcb-4647-ad28-3f9b082392cb" (UID: "20e59f94-8dcb-4647-ad28-3f9b082392cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.621511 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key\") pod \"20e59f94-8dcb-4647-ad28-3f9b082392cb\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.621568 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data\") pod \"20e59f94-8dcb-4647-ad28-3f9b082392cb\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.621602 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdpfr\" (UniqueName: \"kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr\") pod \"20e59f94-8dcb-4647-ad28-3f9b082392cb\" (UID: \"20e59f94-8dcb-4647-ad28-3f9b082392cb\") " Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622410 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622429 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622437 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622446 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28msk\" (UniqueName: \"kubernetes.io/projected/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-kube-api-access-28msk\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622456 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.622464 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20e59f94-8dcb-4647-ad28-3f9b082392cb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.625055 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr" (OuterVolumeSpecName: "kube-api-access-vdpfr") pod "20e59f94-8dcb-4647-ad28-3f9b082392cb" (UID: "20e59f94-8dcb-4647-ad28-3f9b082392cb"). InnerVolumeSpecName "kube-api-access-vdpfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.625387 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "20e59f94-8dcb-4647-ad28-3f9b082392cb" (UID: "20e59f94-8dcb-4647-ad28-3f9b082392cb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.650899 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data" (OuterVolumeSpecName: "config-data") pod "20e59f94-8dcb-4647-ad28-3f9b082392cb" (UID: "20e59f94-8dcb-4647-ad28-3f9b082392cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.663166 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts" (OuterVolumeSpecName: "scripts") pod "20e59f94-8dcb-4647-ad28-3f9b082392cb" (UID: "20e59f94-8dcb-4647-ad28-3f9b082392cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.724762 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20e59f94-8dcb-4647-ad28-3f9b082392cb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.725145 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.725157 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdpfr\" (UniqueName: \"kubernetes.io/projected/20e59f94-8dcb-4647-ad28-3f9b082392cb-kube-api-access-vdpfr\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.725168 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20e59f94-8dcb-4647-ad28-3f9b082392cb-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.752994 5018 scope.go:117] "RemoveContainer" containerID="ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.770130 5018 scope.go:117] "RemoveContainer" containerID="0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" Oct 14 08:55:11 crc kubenswrapper[5018]: E1014 08:55:11.770435 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc\": container with ID starting with 0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc not found: ID does not exist" containerID="0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.770492 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc"} err="failed to get container status \"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc\": rpc error: code = NotFound desc = could not find container \"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc\": container with ID starting with 0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.770522 5018 scope.go:117] "RemoveContainer" containerID="ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" Oct 14 08:55:11 crc kubenswrapper[5018]: E1014 08:55:11.770911 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c\": container with ID starting with ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c not found: ID does not exist" containerID="ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.770935 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c"} err="failed to get container status \"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c\": rpc error: code = NotFound desc = could not find container \"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c\": container with ID starting with ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.770948 5018 scope.go:117] "RemoveContainer" containerID="0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.771399 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc"} err="failed to get container status \"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc\": rpc error: code = NotFound desc = could not find container \"0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc\": container with ID starting with 0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.771418 5018 scope.go:117] "RemoveContainer" containerID="ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.771608 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c"} err="failed to get container status \"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c\": rpc error: code = NotFound desc = could not find container \"ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c\": container with ID starting with ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c not found: ID does not exist" Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.809627 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:55:11 crc kubenswrapper[5018]: I1014 08:55:11.815887 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67ff97b49c-79fh8"] Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.054007 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-z5s89"] Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.067393 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-z5s89"] Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.486433 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-789f4f677c-vvw6r" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.486433 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-789f4f677c-vvw6r" event={"ID":"20e59f94-8dcb-4647-ad28-3f9b082392cb","Type":"ContainerDied","Data":"869aafa8b604261256e436db7b6ff1db2f4968e8a4b30150a818ed6cb57103b7"} Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.486569 5018 scope.go:117] "RemoveContainer" containerID="ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.503722 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.115:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8443: connect: connection refused" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.536408 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.548243 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-789f4f677c-vvw6r"] Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.645460 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" path="/var/lib/kubelet/pods/20e59f94-8dcb-4647-ad28-3f9b082392cb/volumes" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.646285 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b224e5-b8be-4cfa-9ec0-7796ae74e556" path="/var/lib/kubelet/pods/41b224e5-b8be-4cfa-9ec0-7796ae74e556/volumes" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.646967 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" path="/var/lib/kubelet/pods/9b5eed1f-c3f5-4a14-bf72-aaa3338dd974/volumes" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.648212 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" path="/var/lib/kubelet/pods/dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f/volumes" Oct 14 08:55:12 crc kubenswrapper[5018]: I1014 08:55:12.698458 5018 scope.go:117] "RemoveContainer" containerID="0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54" Oct 14 08:55:22 crc kubenswrapper[5018]: I1014 08:55:22.504166 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.115:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8443: connect: connection refused" Oct 14 08:55:32 crc kubenswrapper[5018]: I1014 08:55:32.463232 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:55:32 crc kubenswrapper[5018]: I1014 08:55:32.463824 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:55:32 crc kubenswrapper[5018]: I1014 08:55:32.504114 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-99549d576-lsj55" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.115:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8443: connect: connection refused" Oct 14 08:55:32 crc kubenswrapper[5018]: I1014 08:55:32.504263 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:55:37 crc kubenswrapper[5018]: E1014 08:55:37.717121 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f090ce_afb6_4d48_bf6e_3ad593315c84.slice/crio-conmon-7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcafdd02_670f_4a2b_9e85_7ca0d2da1c3f.slice/crio-d08aaa102ef49015f622f49d0c6f2b9b63a878402b2e757e62c157cb81abf7df\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f090ce_afb6_4d48_bf6e_3ad593315c84.slice/crio-7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-conmon-0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice/crio-ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-conmon-ffbadb2cb561abaaf2ecd33a6aa162632277149a6bb6e4ebb3e4dd53e4d6805c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice/crio-conmon-ee6db04e8cbf0be66298cd4a214e5734f97410e7cf84c2c0377abbe81a5a04e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice/crio-869aafa8b604261256e436db7b6ff1db2f4968e8a4b30150a818ed6cb57103b7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-944d9b4a42daec06b15be9981a6c214376abce4474807485dc2d0b1c53982266\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcafdd02_670f_4a2b_9e85_7ca0d2da1c3f.slice/crio-conmon-7dde28a2c1da8a2574ca2aa7c5dc9f4599270d6c677de4ead24682a4a0120b7c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice/crio-conmon-0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20e59f94_8dcb_4647_ad28_3f9b082392cb.slice/crio-0e0b0c38ef72e1caa8d3ec4a5e26962d30dc38e37641a2a6137e743ca28a6b54.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f090ce_afb6_4d48_bf6e_3ad593315c84.slice/crio-7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcafdd02_670f_4a2b_9e85_7ca0d2da1c3f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5eed1f_c3f5_4a14_bf72_aaa3338dd974.slice/crio-0bc5dc5b9afcd2dde34330ecdb444054c5a7be6954fa13034d68edbaf8240abc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f090ce_afb6_4d48_bf6e_3ad593315c84.slice/crio-conmon-7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.848730 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.859030 5018 generic.go:334] "Generic (PLEG): container finished" podID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerID="7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb" exitCode=137 Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.859076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerDied","Data":"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb"} Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.859108 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-99549d576-lsj55" event={"ID":"d6f090ce-afb6-4d48-bf6e-3ad593315c84","Type":"ContainerDied","Data":"e0f32af6a386824c1dde631598991f5cde7299c0b834e59dc61863f74975748d"} Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.859130 5018 scope.go:117] "RemoveContainer" containerID="7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.859276 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-99549d576-lsj55" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.906444 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.906573 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-879tj\" (UniqueName: \"kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.906771 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.906878 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.906938 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.907001 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.907094 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs\") pod \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\" (UID: \"d6f090ce-afb6-4d48-bf6e-3ad593315c84\") " Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.908220 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs" (OuterVolumeSpecName: "logs") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.908888 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f090ce-afb6-4d48-bf6e-3ad593315c84-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.913015 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj" (OuterVolumeSpecName: "kube-api-access-879tj") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "kube-api-access-879tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.917943 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.949926 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data" (OuterVolumeSpecName: "config-data") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.950036 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.950782 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts" (OuterVolumeSpecName: "scripts") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:37 crc kubenswrapper[5018]: I1014 08:55:37.978498 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d6f090ce-afb6-4d48-bf6e-3ad593315c84" (UID: "d6f090ce-afb6-4d48-bf6e-3ad593315c84"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010178 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010213 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-879tj\" (UniqueName: \"kubernetes.io/projected/d6f090ce-afb6-4d48-bf6e-3ad593315c84-kube-api-access-879tj\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010230 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010242 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010256 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6f090ce-afb6-4d48-bf6e-3ad593315c84-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.010267 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f090ce-afb6-4d48-bf6e-3ad593315c84-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.077939 5018 scope.go:117] "RemoveContainer" containerID="7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.104907 5018 scope.go:117] "RemoveContainer" containerID="7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc" Oct 14 08:55:38 crc kubenswrapper[5018]: E1014 08:55:38.106136 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc\": container with ID starting with 7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc not found: ID does not exist" containerID="7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.106336 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc"} err="failed to get container status \"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc\": rpc error: code = NotFound desc = could not find container \"7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc\": container with ID starting with 7ed2abed6c61cf1bb3f3daee52a621e9335ce582dd90aa60e6523a21d4dd82dc not found: ID does not exist" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.106452 5018 scope.go:117] "RemoveContainer" containerID="7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb" Oct 14 08:55:38 crc kubenswrapper[5018]: E1014 08:55:38.106881 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb\": container with ID starting with 7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb not found: ID does not exist" containerID="7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.106903 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb"} err="failed to get container status \"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb\": rpc error: code = NotFound desc = could not find container \"7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb\": container with ID starting with 7b30ef293c9ac05ed8ac81c906f76fc54a87f6a2ce4e1c098cca03fbcca5bedb not found: ID does not exist" Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.206887 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.214311 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-99549d576-lsj55"] Oct 14 08:55:38 crc kubenswrapper[5018]: I1014 08:55:38.623106 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" path="/var/lib/kubelet/pods/d6f090ce-afb6-4d48-bf6e-3ad593315c84/volumes" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.117610 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.119577 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.119706 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.119780 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.119846 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.119949 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="extract-utilities" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120020 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="extract-utilities" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120098 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120162 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120241 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120311 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120404 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120468 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120544 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="registry-server" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120614 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="registry-server" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120703 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120771 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: E1014 08:55:48.120852 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="extract-content" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.120918 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="extract-content" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121218 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121301 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121383 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f090ce-afb6-4d48-bf6e-3ad593315c84" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121461 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5eed1f-c3f5-4a14-bf72-aaa3338dd974" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121536 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon-log" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121609 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcafdd02-670f-4a2b-9e85-7ca0d2da1c3f" containerName="registry-server" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.121712 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e59f94-8dcb-4647-ad28-3f9b082392cb" containerName="horizon" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.122979 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.160228 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235328 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235412 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235474 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235494 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44lm5\" (UniqueName: \"kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235553 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235576 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.235639 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338167 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338223 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338252 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338399 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338425 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338458 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.338483 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44lm5\" (UniqueName: \"kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.339521 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.339908 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.340001 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.343864 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.344235 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.353483 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.357941 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44lm5\" (UniqueName: \"kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5\") pod \"horizon-74b99bc44-qsr8p\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.450337 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:48 crc kubenswrapper[5018]: I1014 08:55:48.914799 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.012215 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerStarted","Data":"fc66057e4599459dbcba725f4871e513169b148b99ab3a0f13948e5fb286a3f2"} Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.437202 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-kkfbd"] Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.438661 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.448647 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kkfbd"] Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.560102 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7tc5\" (UniqueName: \"kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5\") pod \"heat-db-create-kkfbd\" (UID: \"d0482650-95be-459a-901b-ee5ff6cb9e09\") " pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.663014 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7tc5\" (UniqueName: \"kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5\") pod \"heat-db-create-kkfbd\" (UID: \"d0482650-95be-459a-901b-ee5ff6cb9e09\") " pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.681315 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7tc5\" (UniqueName: \"kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5\") pod \"heat-db-create-kkfbd\" (UID: \"d0482650-95be-459a-901b-ee5ff6cb9e09\") " pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:49 crc kubenswrapper[5018]: I1014 08:55:49.763474 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:50 crc kubenswrapper[5018]: I1014 08:55:50.039229 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerStarted","Data":"392ee2ce4169eab5c3bfa508967090781008c6bb4b91e91bc63269bdd41562f3"} Oct 14 08:55:50 crc kubenswrapper[5018]: I1014 08:55:50.039292 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerStarted","Data":"dcb994f2a796078d866b79a0686477e41a0d979852b18ac2243d743ecb903f27"} Oct 14 08:55:50 crc kubenswrapper[5018]: I1014 08:55:50.066082 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74b99bc44-qsr8p" podStartSLOduration=2.066058816 podStartE2EDuration="2.066058816s" podCreationTimestamp="2025-10-14 08:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:55:50.058851111 +0000 UTC m=+7566.642897748" watchObservedRunningTime="2025-10-14 08:55:50.066058816 +0000 UTC m=+7566.650105453" Oct 14 08:55:50 crc kubenswrapper[5018]: I1014 08:55:50.242563 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kkfbd"] Oct 14 08:55:51 crc kubenswrapper[5018]: I1014 08:55:51.053156 5018 generic.go:334] "Generic (PLEG): container finished" podID="d0482650-95be-459a-901b-ee5ff6cb9e09" containerID="b85abd63af53151ec3ccb50a2dbc7b82688f80e4c0d1741e99d29d0486758ea1" exitCode=0 Oct 14 08:55:51 crc kubenswrapper[5018]: I1014 08:55:51.053224 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kkfbd" event={"ID":"d0482650-95be-459a-901b-ee5ff6cb9e09","Type":"ContainerDied","Data":"b85abd63af53151ec3ccb50a2dbc7b82688f80e4c0d1741e99d29d0486758ea1"} Oct 14 08:55:51 crc kubenswrapper[5018]: I1014 08:55:51.053307 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kkfbd" event={"ID":"d0482650-95be-459a-901b-ee5ff6cb9e09","Type":"ContainerStarted","Data":"4c31bc8c20f7f514875e173272654136c0da656645b4999688e43b0c742139f8"} Oct 14 08:55:52 crc kubenswrapper[5018]: I1014 08:55:52.489088 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:52 crc kubenswrapper[5018]: I1014 08:55:52.630994 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7tc5\" (UniqueName: \"kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5\") pod \"d0482650-95be-459a-901b-ee5ff6cb9e09\" (UID: \"d0482650-95be-459a-901b-ee5ff6cb9e09\") " Oct 14 08:55:52 crc kubenswrapper[5018]: I1014 08:55:52.638015 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5" (OuterVolumeSpecName: "kube-api-access-m7tc5") pod "d0482650-95be-459a-901b-ee5ff6cb9e09" (UID: "d0482650-95be-459a-901b-ee5ff6cb9e09"). InnerVolumeSpecName "kube-api-access-m7tc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:52 crc kubenswrapper[5018]: I1014 08:55:52.734496 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7tc5\" (UniqueName: \"kubernetes.io/projected/d0482650-95be-459a-901b-ee5ff6cb9e09-kube-api-access-m7tc5\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:53 crc kubenswrapper[5018]: I1014 08:55:53.069719 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6p8hr"] Oct 14 08:55:53 crc kubenswrapper[5018]: I1014 08:55:53.085051 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6p8hr"] Oct 14 08:55:53 crc kubenswrapper[5018]: I1014 08:55:53.091309 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kkfbd" event={"ID":"d0482650-95be-459a-901b-ee5ff6cb9e09","Type":"ContainerDied","Data":"4c31bc8c20f7f514875e173272654136c0da656645b4999688e43b0c742139f8"} Oct 14 08:55:53 crc kubenswrapper[5018]: I1014 08:55:53.091369 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c31bc8c20f7f514875e173272654136c0da656645b4999688e43b0c742139f8" Oct 14 08:55:53 crc kubenswrapper[5018]: I1014 08:55:53.091399 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kkfbd" Oct 14 08:55:54 crc kubenswrapper[5018]: I1014 08:55:54.626901 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a812d6-06cc-494e-8862-d3d5a4cdef85" path="/var/lib/kubelet/pods/77a812d6-06cc-494e-8862-d3d5a4cdef85/volumes" Oct 14 08:55:58 crc kubenswrapper[5018]: I1014 08:55:58.451973 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:58 crc kubenswrapper[5018]: I1014 08:55:58.453896 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.617678 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-d90b-account-create-lgjcg"] Oct 14 08:55:59 crc kubenswrapper[5018]: E1014 08:55:59.618910 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0482650-95be-459a-901b-ee5ff6cb9e09" containerName="mariadb-database-create" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.618934 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0482650-95be-459a-901b-ee5ff6cb9e09" containerName="mariadb-database-create" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.619293 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0482650-95be-459a-901b-ee5ff6cb9e09" containerName="mariadb-database-create" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.620353 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.624261 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.649054 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d90b-account-create-lgjcg"] Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.809581 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l76wv\" (UniqueName: \"kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv\") pod \"heat-d90b-account-create-lgjcg\" (UID: \"db6927f2-a17e-4f85-be65-ea18d06166f5\") " pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.911922 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l76wv\" (UniqueName: \"kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv\") pod \"heat-d90b-account-create-lgjcg\" (UID: \"db6927f2-a17e-4f85-be65-ea18d06166f5\") " pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.938528 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l76wv\" (UniqueName: \"kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv\") pod \"heat-d90b-account-create-lgjcg\" (UID: \"db6927f2-a17e-4f85-be65-ea18d06166f5\") " pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:55:59 crc kubenswrapper[5018]: I1014 08:55:59.955890 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:56:00 crc kubenswrapper[5018]: I1014 08:56:00.241449 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d90b-account-create-lgjcg"] Oct 14 08:56:01 crc kubenswrapper[5018]: I1014 08:56:01.204102 5018 generic.go:334] "Generic (PLEG): container finished" podID="db6927f2-a17e-4f85-be65-ea18d06166f5" containerID="e75f3722943cc4d4b297dc5162439acb46bfe1e62a4291fce3f2582331b91793" exitCode=0 Oct 14 08:56:01 crc kubenswrapper[5018]: I1014 08:56:01.204325 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d90b-account-create-lgjcg" event={"ID":"db6927f2-a17e-4f85-be65-ea18d06166f5","Type":"ContainerDied","Data":"e75f3722943cc4d4b297dc5162439acb46bfe1e62a4291fce3f2582331b91793"} Oct 14 08:56:01 crc kubenswrapper[5018]: I1014 08:56:01.204652 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d90b-account-create-lgjcg" event={"ID":"db6927f2-a17e-4f85-be65-ea18d06166f5","Type":"ContainerStarted","Data":"b52e73eaacd717cb9dae18178aeb9dad6041962e9a59cf6f5b45579ec7ba68c3"} Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.464471 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.464996 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.465082 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.466464 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.466590 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e" gracePeriod=600 Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.771965 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.885881 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l76wv\" (UniqueName: \"kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv\") pod \"db6927f2-a17e-4f85-be65-ea18d06166f5\" (UID: \"db6927f2-a17e-4f85-be65-ea18d06166f5\") " Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.892044 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv" (OuterVolumeSpecName: "kube-api-access-l76wv") pod "db6927f2-a17e-4f85-be65-ea18d06166f5" (UID: "db6927f2-a17e-4f85-be65-ea18d06166f5"). InnerVolumeSpecName "kube-api-access-l76wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:02 crc kubenswrapper[5018]: I1014 08:56:02.989309 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l76wv\" (UniqueName: \"kubernetes.io/projected/db6927f2-a17e-4f85-be65-ea18d06166f5-kube-api-access-l76wv\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.050803 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d3e5-account-create-pjcwj"] Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.065640 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d3e5-account-create-pjcwj"] Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.232765 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d90b-account-create-lgjcg" event={"ID":"db6927f2-a17e-4f85-be65-ea18d06166f5","Type":"ContainerDied","Data":"b52e73eaacd717cb9dae18178aeb9dad6041962e9a59cf6f5b45579ec7ba68c3"} Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.232836 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b52e73eaacd717cb9dae18178aeb9dad6041962e9a59cf6f5b45579ec7ba68c3" Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.232882 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d90b-account-create-lgjcg" Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.237363 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e" exitCode=0 Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.237416 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e"} Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.237746 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef"} Oct 14 08:56:03 crc kubenswrapper[5018]: I1014 08:56:03.237805 5018 scope.go:117] "RemoveContainer" containerID="9ef841ae90db78da951994bfa63cc6e41b71935be5e27cb61dc71f760cbf4c44" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.626698 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9e6779-e634-4ca2-8724-4b3962dac90b" path="/var/lib/kubelet/pods/de9e6779-e634-4ca2-8724-4b3962dac90b/volumes" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.694907 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-6ssgc"] Oct 14 08:56:04 crc kubenswrapper[5018]: E1014 08:56:04.695392 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6927f2-a17e-4f85-be65-ea18d06166f5" containerName="mariadb-account-create" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.695410 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6927f2-a17e-4f85-be65-ea18d06166f5" containerName="mariadb-account-create" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.695672 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6927f2-a17e-4f85-be65-ea18d06166f5" containerName="mariadb-account-create" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.696416 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.699401 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7k8sb" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.700052 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.708671 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6ssgc"] Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.829763 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckjpf\" (UniqueName: \"kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.829848 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.829889 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.932022 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckjpf\" (UniqueName: \"kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.932096 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.932143 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.937459 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.938372 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:04 crc kubenswrapper[5018]: I1014 08:56:04.961065 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckjpf\" (UniqueName: \"kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf\") pod \"heat-db-sync-6ssgc\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:05 crc kubenswrapper[5018]: I1014 08:56:05.018540 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:05 crc kubenswrapper[5018]: W1014 08:56:05.493917 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb07a1347_0c41_470f_b4ce_d916787935c0.slice/crio-02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8 WatchSource:0}: Error finding container 02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8: Status 404 returned error can't find the container with id 02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8 Oct 14 08:56:05 crc kubenswrapper[5018]: I1014 08:56:05.497136 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6ssgc"] Oct 14 08:56:06 crc kubenswrapper[5018]: I1014 08:56:06.273559 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ssgc" event={"ID":"b07a1347-0c41-470f-b4ce-d916787935c0","Type":"ContainerStarted","Data":"02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8"} Oct 14 08:56:09 crc kubenswrapper[5018]: I1014 08:56:09.724544 5018 scope.go:117] "RemoveContainer" containerID="47f7114dfd22fca5d153892b0fddff620b178f6206ed69f9ff4e8c54b7a1b3ed" Oct 14 08:56:10 crc kubenswrapper[5018]: I1014 08:56:10.179408 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:56:11 crc kubenswrapper[5018]: I1014 08:56:11.747962 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 08:56:11 crc kubenswrapper[5018]: I1014 08:56:11.819488 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:56:11 crc kubenswrapper[5018]: I1014 08:56:11.819742 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon-log" containerID="cri-o://31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1" gracePeriod=30 Oct 14 08:56:11 crc kubenswrapper[5018]: I1014 08:56:11.819863 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" containerID="cri-o://cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41" gracePeriod=30 Oct 14 08:56:12 crc kubenswrapper[5018]: I1014 08:56:12.031345 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-6l49x"] Oct 14 08:56:12 crc kubenswrapper[5018]: I1014 08:56:12.045477 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-6l49x"] Oct 14 08:56:12 crc kubenswrapper[5018]: I1014 08:56:12.621000 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9024d8b2-89ef-4304-9eb7-5d684864402c" path="/var/lib/kubelet/pods/9024d8b2-89ef-4304-9eb7-5d684864402c/volumes" Oct 14 08:56:13 crc kubenswrapper[5018]: I1014 08:56:13.706910 5018 scope.go:117] "RemoveContainer" containerID="41cc3322c56458de6f5bcf72edcc9e797ada50be777ccec9a283724dbd865ee8" Oct 14 08:56:14 crc kubenswrapper[5018]: I1014 08:56:14.020603 5018 scope.go:117] "RemoveContainer" containerID="57f99d2fd42f8b9bef5c70891072754cba8a507005f9e58d49af669989298c89" Oct 14 08:56:14 crc kubenswrapper[5018]: I1014 08:56:14.075228 5018 scope.go:117] "RemoveContainer" containerID="b03117968d08499e531fc7721c3ce7ae44a008e802626016b03f23faa1b4b08f" Oct 14 08:56:14 crc kubenswrapper[5018]: I1014 08:56:14.255086 5018 scope.go:117] "RemoveContainer" containerID="ed39e92cd28c36a04d30e0fa9f9aa9bd7b1e722ce24684bfbdf53b76ada30c21" Oct 14 08:56:14 crc kubenswrapper[5018]: I1014 08:56:14.311005 5018 scope.go:117] "RemoveContainer" containerID="295a83f13373ddcea2ff47866b73fdb6d1f09e67500d56c6ab48553e5df24d71" Oct 14 08:56:14 crc kubenswrapper[5018]: I1014 08:56:14.977991 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.116:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:58342->10.217.1.116:8443: read: connection reset by peer" Oct 14 08:56:15 crc kubenswrapper[5018]: I1014 08:56:15.387062 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ssgc" event={"ID":"b07a1347-0c41-470f-b4ce-d916787935c0","Type":"ContainerStarted","Data":"815430d7601a06039490b1146f9c8d9aafce21bf0b6517a403372225de85d41d"} Oct 14 08:56:15 crc kubenswrapper[5018]: I1014 08:56:15.390788 5018 generic.go:334] "Generic (PLEG): container finished" podID="1e10a610-e422-4785-89a4-5c1beb50c116" containerID="cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41" exitCode=0 Oct 14 08:56:15 crc kubenswrapper[5018]: I1014 08:56:15.390828 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerDied","Data":"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41"} Oct 14 08:56:15 crc kubenswrapper[5018]: I1014 08:56:15.404195 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-6ssgc" podStartSLOduration=2.8045043549999997 podStartE2EDuration="11.40417501s" podCreationTimestamp="2025-10-14 08:56:04 +0000 UTC" firstStartedPulling="2025-10-14 08:56:05.496358703 +0000 UTC m=+7582.080405330" lastFinishedPulling="2025-10-14 08:56:14.096029358 +0000 UTC m=+7590.680075985" observedRunningTime="2025-10-14 08:56:15.402034429 +0000 UTC m=+7591.986081056" watchObservedRunningTime="2025-10-14 08:56:15.40417501 +0000 UTC m=+7591.988221637" Oct 14 08:56:17 crc kubenswrapper[5018]: I1014 08:56:17.417071 5018 generic.go:334] "Generic (PLEG): container finished" podID="b07a1347-0c41-470f-b4ce-d916787935c0" containerID="815430d7601a06039490b1146f9c8d9aafce21bf0b6517a403372225de85d41d" exitCode=0 Oct 14 08:56:17 crc kubenswrapper[5018]: I1014 08:56:17.417192 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ssgc" event={"ID":"b07a1347-0c41-470f-b4ce-d916787935c0","Type":"ContainerDied","Data":"815430d7601a06039490b1146f9c8d9aafce21bf0b6517a403372225de85d41d"} Oct 14 08:56:18 crc kubenswrapper[5018]: I1014 08:56:18.869229 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.007452 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data\") pod \"b07a1347-0c41-470f-b4ce-d916787935c0\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.007671 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckjpf\" (UniqueName: \"kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf\") pod \"b07a1347-0c41-470f-b4ce-d916787935c0\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.007850 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle\") pod \"b07a1347-0c41-470f-b4ce-d916787935c0\" (UID: \"b07a1347-0c41-470f-b4ce-d916787935c0\") " Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.015222 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf" (OuterVolumeSpecName: "kube-api-access-ckjpf") pod "b07a1347-0c41-470f-b4ce-d916787935c0" (UID: "b07a1347-0c41-470f-b4ce-d916787935c0"). InnerVolumeSpecName "kube-api-access-ckjpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.038890 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b07a1347-0c41-470f-b4ce-d916787935c0" (UID: "b07a1347-0c41-470f-b4ce-d916787935c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.110550 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.110590 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckjpf\" (UniqueName: \"kubernetes.io/projected/b07a1347-0c41-470f-b4ce-d916787935c0-kube-api-access-ckjpf\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.116285 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data" (OuterVolumeSpecName: "config-data") pod "b07a1347-0c41-470f-b4ce-d916787935c0" (UID: "b07a1347-0c41-470f-b4ce-d916787935c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.212022 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07a1347-0c41-470f-b4ce-d916787935c0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.466136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ssgc" event={"ID":"b07a1347-0c41-470f-b4ce-d916787935c0","Type":"ContainerDied","Data":"02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8"} Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.466496 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02aadeb357fe1db1d7655dd4cd5971cc4eae75916788990439a43ac1c718e8d8" Oct 14 08:56:19 crc kubenswrapper[5018]: I1014 08:56:19.466186 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ssgc" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.677539 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:56:20 crc kubenswrapper[5018]: E1014 08:56:20.678180 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b07a1347-0c41-470f-b4ce-d916787935c0" containerName="heat-db-sync" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.678198 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b07a1347-0c41-470f-b4ce-d916787935c0" containerName="heat-db-sync" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.678411 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b07a1347-0c41-470f-b4ce-d916787935c0" containerName="heat-db-sync" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.679211 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.681759 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.682159 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7k8sb" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.684071 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.758773 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.768738 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.768807 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.768893 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9f6g\" (UniqueName: \"kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.769155 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.811612 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.812872 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.818002 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.819234 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.819503 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.822800 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.834268 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.841737 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.874233 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.874392 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.874447 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.874496 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9f6g\" (UniqueName: \"kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.882711 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.884533 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.889426 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9f6g\" (UniqueName: \"kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.906427 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle\") pod \"heat-engine-55c7b9d586-sxknt\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.975817 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976279 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgcnn\" (UniqueName: \"kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976314 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976359 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976529 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976568 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976613 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:20 crc kubenswrapper[5018]: I1014 08:56:20.976673 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbqc6\" (UniqueName: \"kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.011490 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.083951 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084158 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084191 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084234 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbqc6\" (UniqueName: \"kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084331 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084480 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgcnn\" (UniqueName: \"kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.084515 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.088176 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.089409 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.091123 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.091522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.092926 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.107687 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbqc6\" (UniqueName: \"kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.108385 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle\") pod \"heat-cfnapi-7c78564b6-p8jl2\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.112541 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgcnn\" (UniqueName: \"kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn\") pod \"heat-api-687546b8cd-dpj5m\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.140155 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.148335 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.550174 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.658047 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:21 crc kubenswrapper[5018]: I1014 08:56:21.743796 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:21 crc kubenswrapper[5018]: W1014 08:56:21.746033 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod267cbd65_9ace_4760_8602_3adb5e589748.slice/crio-1aed9d16b90ee0deeb4202ca2565fe68d2a2b721e051a2c394385b685244ae74 WatchSource:0}: Error finding container 1aed9d16b90ee0deeb4202ca2565fe68d2a2b721e051a2c394385b685244ae74: Status 404 returned error can't find the container with id 1aed9d16b90ee0deeb4202ca2565fe68d2a2b721e051a2c394385b685244ae74 Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.496103 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-687546b8cd-dpj5m" event={"ID":"267cbd65-9ace-4760-8602-3adb5e589748","Type":"ContainerStarted","Data":"1aed9d16b90ee0deeb4202ca2565fe68d2a2b721e051a2c394385b685244ae74"} Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.497594 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" event={"ID":"1eb4912b-5bce-41ac-a728-aa0bdaea70c3","Type":"ContainerStarted","Data":"0914748935f80dd53e4f09396530f025ee4ecc409387ab3f01a9e8cf8f861a5e"} Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.498966 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c7b9d586-sxknt" event={"ID":"8bb51c74-168b-4420-ba0b-b2ce254929d8","Type":"ContainerStarted","Data":"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7"} Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.498987 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c7b9d586-sxknt" event={"ID":"8bb51c74-168b-4420-ba0b-b2ce254929d8","Type":"ContainerStarted","Data":"d443e735fb983688f946c7ced2ed36524c308ab86ccdb6e6d8ca2b519aa265d1"} Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.500192 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.522157 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55c7b9d586-sxknt" podStartSLOduration=2.522142969 podStartE2EDuration="2.522142969s" podCreationTimestamp="2025-10-14 08:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:22.518274119 +0000 UTC m=+7599.102320746" watchObservedRunningTime="2025-10-14 08:56:22.522142969 +0000 UTC m=+7599.106189596" Oct 14 08:56:22 crc kubenswrapper[5018]: I1014 08:56:22.562595 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.116:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8443: connect: connection refused" Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.526915 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-687546b8cd-dpj5m" event={"ID":"267cbd65-9ace-4760-8602-3adb5e589748","Type":"ContainerStarted","Data":"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a"} Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.527384 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.529443 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" event={"ID":"1eb4912b-5bce-41ac-a728-aa0bdaea70c3","Type":"ContainerStarted","Data":"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a"} Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.529490 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.547039 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-687546b8cd-dpj5m" podStartSLOduration=2.31844039 podStartE2EDuration="3.54702005s" podCreationTimestamp="2025-10-14 08:56:20 +0000 UTC" firstStartedPulling="2025-10-14 08:56:21.74815971 +0000 UTC m=+7598.332206337" lastFinishedPulling="2025-10-14 08:56:22.97673937 +0000 UTC m=+7599.560785997" observedRunningTime="2025-10-14 08:56:23.540996209 +0000 UTC m=+7600.125042836" watchObservedRunningTime="2025-10-14 08:56:23.54702005 +0000 UTC m=+7600.131066677" Oct 14 08:56:23 crc kubenswrapper[5018]: I1014 08:56:23.570236 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" podStartSLOduration=2.2782947780000002 podStartE2EDuration="3.570219089s" podCreationTimestamp="2025-10-14 08:56:20 +0000 UTC" firstStartedPulling="2025-10-14 08:56:21.66653544 +0000 UTC m=+7598.250582067" lastFinishedPulling="2025-10-14 08:56:22.958459731 +0000 UTC m=+7599.542506378" observedRunningTime="2025-10-14 08:56:23.563371175 +0000 UTC m=+7600.147417812" watchObservedRunningTime="2025-10-14 08:56:23.570219089 +0000 UTC m=+7600.154265716" Oct 14 08:56:27 crc kubenswrapper[5018]: I1014 08:56:27.961163 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 08:56:27 crc kubenswrapper[5018]: I1014 08:56:27.963043 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:27.993843 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.095554 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.096843 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.111834 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.113162 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.126426 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.127093 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt9x7\" (UniqueName: \"kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.127274 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.127451 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.127561 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.137500 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.228935 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.228989 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdbnf\" (UniqueName: \"kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229023 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4tl4\" (UniqueName: \"kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229055 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229085 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229172 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229336 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229380 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229438 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229472 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229597 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.229760 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt9x7\" (UniqueName: \"kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.240646 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.241522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.248515 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.250601 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt9x7\" (UniqueName: \"kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7\") pod \"heat-engine-67b5f9ffc9-8mcjc\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331341 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdbnf\" (UniqueName: \"kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331420 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4tl4\" (UniqueName: \"kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331459 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331495 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331550 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331698 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331737 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.331815 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.336745 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.337124 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.337258 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.337338 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.337714 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.339040 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.339259 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.351949 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdbnf\" (UniqueName: \"kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf\") pod \"heat-api-5697d4c49-csh9w\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.353264 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4tl4\" (UniqueName: \"kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4\") pod \"heat-cfnapi-59f5d7968c-jh7qb\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.414243 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.453515 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.837192 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.894806 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:28 crc kubenswrapper[5018]: W1014 08:56:28.895952 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58631f04_ebd4_426d_bd77_98b5ff6e00f9.slice/crio-f243c35f9b3a225d6728ef581eed2da61d3f2243ec166fff21105741b07f38a5 WatchSource:0}: Error finding container f243c35f9b3a225d6728ef581eed2da61d3f2243ec166fff21105741b07f38a5: Status 404 returned error can't find the container with id f243c35f9b3a225d6728ef581eed2da61d3f2243ec166fff21105741b07f38a5 Oct 14 08:56:28 crc kubenswrapper[5018]: W1014 08:56:28.947392 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod003d6dc1_3777_4d37_a709_55a2d6c11623.slice/crio-d811abd09dc823eb068dcb0c3a764af89b9a66023352a9c45b0168fd30041ba2 WatchSource:0}: Error finding container d811abd09dc823eb068dcb0c3a764af89b9a66023352a9c45b0168fd30041ba2: Status 404 returned error can't find the container with id d811abd09dc823eb068dcb0c3a764af89b9a66023352a9c45b0168fd30041ba2 Oct 14 08:56:28 crc kubenswrapper[5018]: I1014 08:56:28.952747 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.331299 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.331809 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" containerID="cri-o://b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a" gracePeriod=60 Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.341696 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.341941 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-687546b8cd-dpj5m" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" containerID="cri-o://5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a" gracePeriod=60 Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.344509 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.126:8000/healthcheck\": EOF" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.363872 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.365168 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.367971 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.368500 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.378052 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.379706 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.385365 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.385495 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.389538 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.429378 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.447738 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-687546b8cd-dpj5m" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.125:8004/healthcheck\": EOF" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.556967 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557054 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557294 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557396 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557437 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557571 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cst47\" (UniqueName: \"kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557725 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557853 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.557999 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.558028 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.558093 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmvf7\" (UniqueName: \"kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.558162 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.639092 5018 generic.go:334] "Generic (PLEG): container finished" podID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerID="a739416dfb0365124facfda37c34a9522ef45a62144836fc45b91b1a176d02b0" exitCode=1 Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.639162 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" event={"ID":"003d6dc1-3777-4d37-a709-55a2d6c11623","Type":"ContainerDied","Data":"a739416dfb0365124facfda37c34a9522ef45a62144836fc45b91b1a176d02b0"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.639188 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" event={"ID":"003d6dc1-3777-4d37-a709-55a2d6c11623","Type":"ContainerStarted","Data":"d811abd09dc823eb068dcb0c3a764af89b9a66023352a9c45b0168fd30041ba2"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.640488 5018 scope.go:117] "RemoveContainer" containerID="a739416dfb0365124facfda37c34a9522ef45a62144836fc45b91b1a176d02b0" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.643397 5018 generic.go:334] "Generic (PLEG): container finished" podID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerID="862f35f062f77fcd96f867259ae15606817b3fb3283572180c1d8820f6981834" exitCode=1 Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.643478 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5697d4c49-csh9w" event={"ID":"58631f04-ebd4-426d-bd77-98b5ff6e00f9","Type":"ContainerDied","Data":"862f35f062f77fcd96f867259ae15606817b3fb3283572180c1d8820f6981834"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.643516 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5697d4c49-csh9w" event={"ID":"58631f04-ebd4-426d-bd77-98b5ff6e00f9","Type":"ContainerStarted","Data":"f243c35f9b3a225d6728ef581eed2da61d3f2243ec166fff21105741b07f38a5"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.644319 5018 scope.go:117] "RemoveContainer" containerID="862f35f062f77fcd96f867259ae15606817b3fb3283572180c1d8820f6981834" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.646731 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" event={"ID":"214ab07d-ba44-499b-bbad-9a9ec59eeaa3","Type":"ContainerStarted","Data":"7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.646773 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" event={"ID":"214ab07d-ba44-499b-bbad-9a9ec59eeaa3","Type":"ContainerStarted","Data":"4c8e3e7942f8d8ca1b6a5a662c86ee3404d18a5b685c656a0a9aea0866777d66"} Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.646888 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.662905 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmvf7\" (UniqueName: \"kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.662983 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663015 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663070 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663157 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663191 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663237 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663265 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cst47\" (UniqueName: \"kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663459 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663555 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.663573 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.668354 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.675817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.676183 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.678118 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.679520 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.679932 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.683387 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.686159 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.686719 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.687876 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podStartSLOduration=2.687839124 podStartE2EDuration="2.687839124s" podCreationTimestamp="2025-10-14 08:56:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:29.67610949 +0000 UTC m=+7606.260156117" watchObservedRunningTime="2025-10-14 08:56:29.687839124 +0000 UTC m=+7606.271885751" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.703726 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.710456 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cst47\" (UniqueName: \"kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47\") pod \"heat-api-576557876d-fngj4\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.713537 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmvf7\" (UniqueName: \"kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7\") pod \"heat-cfnapi-76cf75d58b-v66bz\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.982033 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:29 crc kubenswrapper[5018]: I1014 08:56:29.995411 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.419760 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 08:56:30 crc kubenswrapper[5018]: W1014 08:56:30.419906 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod642159cf_6607_450f_84b9_f8ac137d173d.slice/crio-8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc WatchSource:0}: Error finding container 8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc: Status 404 returned error can't find the container with id 8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.495087 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.656485 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-576557876d-fngj4" event={"ID":"eba7274f-a215-4784-a540-06013b44a29c","Type":"ContainerStarted","Data":"d5d5f6908760c77f469d24ae4f12e61f5f9796e2bc3aa135771af4990ad3b26b"} Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.658278 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" event={"ID":"642159cf-6607-450f-84b9-f8ac137d173d","Type":"ContainerStarted","Data":"8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc"} Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.664896 5018 generic.go:334] "Generic (PLEG): container finished" podID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerID="237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9" exitCode=1 Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.665018 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5697d4c49-csh9w" event={"ID":"58631f04-ebd4-426d-bd77-98b5ff6e00f9","Type":"ContainerDied","Data":"237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9"} Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.665052 5018 scope.go:117] "RemoveContainer" containerID="862f35f062f77fcd96f867259ae15606817b3fb3283572180c1d8820f6981834" Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.665567 5018 scope.go:117] "RemoveContainer" containerID="237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9" Oct 14 08:56:30 crc kubenswrapper[5018]: E1014 08:56:30.665900 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5697d4c49-csh9w_openstack(58631f04-ebd4-426d-bd77-98b5ff6e00f9)\"" pod="openstack/heat-api-5697d4c49-csh9w" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.677110 5018 generic.go:334] "Generic (PLEG): container finished" podID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" exitCode=1 Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.678518 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" event={"ID":"003d6dc1-3777-4d37-a709-55a2d6c11623","Type":"ContainerDied","Data":"b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9"} Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.688090 5018 scope.go:117] "RemoveContainer" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" Oct 14 08:56:30 crc kubenswrapper[5018]: E1014 08:56:30.688762 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59f5d7968c-jh7qb_openstack(003d6dc1-3777-4d37-a709-55a2d6c11623)\"" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" Oct 14 08:56:30 crc kubenswrapper[5018]: I1014 08:56:30.714801 5018 scope.go:117] "RemoveContainer" containerID="a739416dfb0365124facfda37c34a9522ef45a62144836fc45b91b1a176d02b0" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.089429 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.687784 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" event={"ID":"642159cf-6607-450f-84b9-f8ac137d173d","Type":"ContainerStarted","Data":"8e5bc391462c1807435f3141698540e3dbc744434e432fffc72f05920e38ffb9"} Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.687946 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.691542 5018 scope.go:117] "RemoveContainer" containerID="237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9" Oct 14 08:56:31 crc kubenswrapper[5018]: E1014 08:56:31.692064 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5697d4c49-csh9w_openstack(58631f04-ebd4-426d-bd77-98b5ff6e00f9)\"" pod="openstack/heat-api-5697d4c49-csh9w" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.696280 5018 scope.go:117] "RemoveContainer" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" Oct 14 08:56:31 crc kubenswrapper[5018]: E1014 08:56:31.696538 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59f5d7968c-jh7qb_openstack(003d6dc1-3777-4d37-a709-55a2d6c11623)\"" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.697130 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-576557876d-fngj4" event={"ID":"eba7274f-a215-4784-a540-06013b44a29c","Type":"ContainerStarted","Data":"c337d9341511431c4a0f5f57944ece931c9e13a68e5f6796f87487029392bc65"} Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.697307 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.704678 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" podStartSLOduration=2.7046348890000003 podStartE2EDuration="2.704634889s" podCreationTimestamp="2025-10-14 08:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:31.703742564 +0000 UTC m=+7608.287789191" watchObservedRunningTime="2025-10-14 08:56:31.704634889 +0000 UTC m=+7608.288681506" Oct 14 08:56:31 crc kubenswrapper[5018]: I1014 08:56:31.733149 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-576557876d-fngj4" podStartSLOduration=2.733129579 podStartE2EDuration="2.733129579s" podCreationTimestamp="2025-10-14 08:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:31.729127695 +0000 UTC m=+7608.313174322" watchObservedRunningTime="2025-10-14 08:56:31.733129579 +0000 UTC m=+7608.317176206" Oct 14 08:56:32 crc kubenswrapper[5018]: I1014 08:56:32.562541 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5cdc789c6-8thlf" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.116:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8443: connect: connection refused" Oct 14 08:56:32 crc kubenswrapper[5018]: I1014 08:56:32.563221 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.414702 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.415014 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.415481 5018 scope.go:117] "RemoveContainer" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" Oct 14 08:56:33 crc kubenswrapper[5018]: E1014 08:56:33.415826 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59f5d7968c-jh7qb_openstack(003d6dc1-3777-4d37-a709-55a2d6c11623)\"" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.453875 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.453934 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.454945 5018 scope.go:117] "RemoveContainer" containerID="237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9" Oct 14 08:56:33 crc kubenswrapper[5018]: E1014 08:56:33.455356 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5697d4c49-csh9w_openstack(58631f04-ebd4-426d-bd77-98b5ff6e00f9)\"" pod="openstack/heat-api-5697d4c49-csh9w" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" Oct 14 08:56:33 crc kubenswrapper[5018]: I1014 08:56:33.716706 5018 scope.go:117] "RemoveContainer" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" Oct 14 08:56:33 crc kubenswrapper[5018]: E1014 08:56:33.716942 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59f5d7968c-jh7qb_openstack(003d6dc1-3777-4d37-a709-55a2d6c11623)\"" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" Oct 14 08:56:34 crc kubenswrapper[5018]: I1014 08:56:34.767417 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-687546b8cd-dpj5m" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.125:8004/healthcheck\": read tcp 10.217.0.2:57150->10.217.1.125:8004: read: connection reset by peer" Oct 14 08:56:34 crc kubenswrapper[5018]: I1014 08:56:34.783803 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.126:8000/healthcheck\": read tcp 10.217.0.2:59528->10.217.1.126:8000: read: connection reset by peer" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.405438 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.411793 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.501966 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom\") pod \"267cbd65-9ace-4760-8602-3adb5e589748\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.502355 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle\") pod \"267cbd65-9ace-4760-8602-3adb5e589748\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.502531 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgcnn\" (UniqueName: \"kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn\") pod \"267cbd65-9ace-4760-8602-3adb5e589748\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.502670 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data\") pod \"267cbd65-9ace-4760-8602-3adb5e589748\" (UID: \"267cbd65-9ace-4760-8602-3adb5e589748\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.507428 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn" (OuterVolumeSpecName: "kube-api-access-pgcnn") pod "267cbd65-9ace-4760-8602-3adb5e589748" (UID: "267cbd65-9ace-4760-8602-3adb5e589748"). InnerVolumeSpecName "kube-api-access-pgcnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.507855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "267cbd65-9ace-4760-8602-3adb5e589748" (UID: "267cbd65-9ace-4760-8602-3adb5e589748"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.533934 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "267cbd65-9ace-4760-8602-3adb5e589748" (UID: "267cbd65-9ace-4760-8602-3adb5e589748"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.581276 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data" (OuterVolumeSpecName: "config-data") pod "267cbd65-9ace-4760-8602-3adb5e589748" (UID: "267cbd65-9ace-4760-8602-3adb5e589748"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.604348 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom\") pod \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.604501 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle\") pod \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.604854 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data\") pod \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.605033 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbqc6\" (UniqueName: \"kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6\") pod \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\" (UID: \"1eb4912b-5bce-41ac-a728-aa0bdaea70c3\") " Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.605802 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.605952 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.606072 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgcnn\" (UniqueName: \"kubernetes.io/projected/267cbd65-9ace-4760-8602-3adb5e589748-kube-api-access-pgcnn\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.606204 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267cbd65-9ace-4760-8602-3adb5e589748-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.607002 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1eb4912b-5bce-41ac-a728-aa0bdaea70c3" (UID: "1eb4912b-5bce-41ac-a728-aa0bdaea70c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.610337 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6" (OuterVolumeSpecName: "kube-api-access-fbqc6") pod "1eb4912b-5bce-41ac-a728-aa0bdaea70c3" (UID: "1eb4912b-5bce-41ac-a728-aa0bdaea70c3"). InnerVolumeSpecName "kube-api-access-fbqc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.635333 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eb4912b-5bce-41ac-a728-aa0bdaea70c3" (UID: "1eb4912b-5bce-41ac-a728-aa0bdaea70c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.657429 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data" (OuterVolumeSpecName: "config-data") pod "1eb4912b-5bce-41ac-a728-aa0bdaea70c3" (UID: "1eb4912b-5bce-41ac-a728-aa0bdaea70c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.709024 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.709077 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbqc6\" (UniqueName: \"kubernetes.io/projected/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-kube-api-access-fbqc6\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.709099 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.709120 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb4912b-5bce-41ac-a728-aa0bdaea70c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.752343 5018 generic.go:334] "Generic (PLEG): container finished" podID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerID="b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a" exitCode=0 Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.752427 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" event={"ID":"1eb4912b-5bce-41ac-a728-aa0bdaea70c3","Type":"ContainerDied","Data":"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a"} Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.752442 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.752482 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c78564b6-p8jl2" event={"ID":"1eb4912b-5bce-41ac-a728-aa0bdaea70c3","Type":"ContainerDied","Data":"0914748935f80dd53e4f09396530f025ee4ecc409387ab3f01a9e8cf8f861a5e"} Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.752508 5018 scope.go:117] "RemoveContainer" containerID="b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.755779 5018 generic.go:334] "Generic (PLEG): container finished" podID="267cbd65-9ace-4760-8602-3adb5e589748" containerID="5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a" exitCode=0 Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.755847 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-687546b8cd-dpj5m" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.755874 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-687546b8cd-dpj5m" event={"ID":"267cbd65-9ace-4760-8602-3adb5e589748","Type":"ContainerDied","Data":"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a"} Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.756538 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-687546b8cd-dpj5m" event={"ID":"267cbd65-9ace-4760-8602-3adb5e589748","Type":"ContainerDied","Data":"1aed9d16b90ee0deeb4202ca2565fe68d2a2b721e051a2c394385b685244ae74"} Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.795162 5018 scope.go:117] "RemoveContainer" containerID="b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a" Oct 14 08:56:35 crc kubenswrapper[5018]: E1014 08:56:35.795837 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a\": container with ID starting with b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a not found: ID does not exist" containerID="b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.795876 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a"} err="failed to get container status \"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a\": rpc error: code = NotFound desc = could not find container \"b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a\": container with ID starting with b962ec0da076e20538391c8b0c24d13d9b2ce62f572f99acc10366dd06dcbe3a not found: ID does not exist" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.795906 5018 scope.go:117] "RemoveContainer" containerID="5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.811964 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.826750 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7c78564b6-p8jl2"] Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.834139 5018 scope.go:117] "RemoveContainer" containerID="5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a" Oct 14 08:56:35 crc kubenswrapper[5018]: E1014 08:56:35.834751 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a\": container with ID starting with 5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a not found: ID does not exist" containerID="5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.834790 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a"} err="failed to get container status \"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a\": rpc error: code = NotFound desc = could not find container \"5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a\": container with ID starting with 5dda9e59f907b54bae689247822f4eb6f5e773d9f6ba002889fe45e17c2cc60a not found: ID does not exist" Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.835148 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:35 crc kubenswrapper[5018]: I1014 08:56:35.847478 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-687546b8cd-dpj5m"] Oct 14 08:56:36 crc kubenswrapper[5018]: I1014 08:56:36.621828 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" path="/var/lib/kubelet/pods/1eb4912b-5bce-41ac-a728-aa0bdaea70c3/volumes" Oct 14 08:56:36 crc kubenswrapper[5018]: I1014 08:56:36.624185 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="267cbd65-9ace-4760-8602-3adb5e589748" path="/var/lib/kubelet/pods/267cbd65-9ace-4760-8602-3adb5e589748/volumes" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.198818 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.239587 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-576557876d-fngj4" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.282692 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.337367 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.731869 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.738922 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.846852 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.846837 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59f5d7968c-jh7qb" event={"ID":"003d6dc1-3777-4d37-a709-55a2d6c11623","Type":"ContainerDied","Data":"d811abd09dc823eb068dcb0c3a764af89b9a66023352a9c45b0168fd30041ba2"} Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.847017 5018 scope.go:117] "RemoveContainer" containerID="b04f10b8c0b7041688d19f34289996d7dc588b3930b417681d9f6da33ad98fb9" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.852199 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5697d4c49-csh9w" event={"ID":"58631f04-ebd4-426d-bd77-98b5ff6e00f9","Type":"ContainerDied","Data":"f243c35f9b3a225d6728ef581eed2da61d3f2243ec166fff21105741b07f38a5"} Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.852300 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5697d4c49-csh9w" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869276 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data\") pod \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869394 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdbnf\" (UniqueName: \"kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf\") pod \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869485 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle\") pod \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4tl4\" (UniqueName: \"kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4\") pod \"003d6dc1-3777-4d37-a709-55a2d6c11623\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869604 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle\") pod \"003d6dc1-3777-4d37-a709-55a2d6c11623\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869667 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom\") pod \"003d6dc1-3777-4d37-a709-55a2d6c11623\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.869696 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data\") pod \"003d6dc1-3777-4d37-a709-55a2d6c11623\" (UID: \"003d6dc1-3777-4d37-a709-55a2d6c11623\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.870720 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom\") pod \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\" (UID: \"58631f04-ebd4-426d-bd77-98b5ff6e00f9\") " Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.875461 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4" (OuterVolumeSpecName: "kube-api-access-t4tl4") pod "003d6dc1-3777-4d37-a709-55a2d6c11623" (UID: "003d6dc1-3777-4d37-a709-55a2d6c11623"). InnerVolumeSpecName "kube-api-access-t4tl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.879116 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "58631f04-ebd4-426d-bd77-98b5ff6e00f9" (UID: "58631f04-ebd4-426d-bd77-98b5ff6e00f9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.879921 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf" (OuterVolumeSpecName: "kube-api-access-gdbnf") pod "58631f04-ebd4-426d-bd77-98b5ff6e00f9" (UID: "58631f04-ebd4-426d-bd77-98b5ff6e00f9"). InnerVolumeSpecName "kube-api-access-gdbnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.896852 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "003d6dc1-3777-4d37-a709-55a2d6c11623" (UID: "003d6dc1-3777-4d37-a709-55a2d6c11623"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.911887 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "003d6dc1-3777-4d37-a709-55a2d6c11623" (UID: "003d6dc1-3777-4d37-a709-55a2d6c11623"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.930486 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58631f04-ebd4-426d-bd77-98b5ff6e00f9" (UID: "58631f04-ebd4-426d-bd77-98b5ff6e00f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.968830 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data" (OuterVolumeSpecName: "config-data") pod "003d6dc1-3777-4d37-a709-55a2d6c11623" (UID: "003d6dc1-3777-4d37-a709-55a2d6c11623"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973067 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973098 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdbnf\" (UniqueName: \"kubernetes.io/projected/58631f04-ebd4-426d-bd77-98b5ff6e00f9-kube-api-access-gdbnf\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973110 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973119 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4tl4\" (UniqueName: \"kubernetes.io/projected/003d6dc1-3777-4d37-a709-55a2d6c11623-kube-api-access-t4tl4\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973127 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973135 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.973143 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/003d6dc1-3777-4d37-a709-55a2d6c11623-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:41 crc kubenswrapper[5018]: I1014 08:56:41.977751 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data" (OuterVolumeSpecName: "config-data") pod "58631f04-ebd4-426d-bd77-98b5ff6e00f9" (UID: "58631f04-ebd4-426d-bd77-98b5ff6e00f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.056150 5018 scope.go:117] "RemoveContainer" containerID="237e2a5c88a7f9d605924d521c62b67bc4514513eb436a8e4ef6f0e2e046bac9" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.075631 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58631f04-ebd4-426d-bd77-98b5ff6e00f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.176339 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.183670 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.195473 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-59f5d7968c-jh7qb"] Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.210967 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.221640 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5697d4c49-csh9w"] Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278532 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dx4z\" (UniqueName: \"kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278614 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278660 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278708 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278789 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278913 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.278986 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle\") pod \"1e10a610-e422-4785-89a4-5c1beb50c116\" (UID: \"1e10a610-e422-4785-89a4-5c1beb50c116\") " Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.279182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs" (OuterVolumeSpecName: "logs") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.279411 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e10a610-e422-4785-89a4-5c1beb50c116-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.281989 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.283518 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z" (OuterVolumeSpecName: "kube-api-access-2dx4z") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "kube-api-access-2dx4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.300459 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts" (OuterVolumeSpecName: "scripts") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.304440 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data" (OuterVolumeSpecName: "config-data") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.308764 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.361577 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "1e10a610-e422-4785-89a4-5c1beb50c116" (UID: "1e10a610-e422-4785-89a4-5c1beb50c116"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384135 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384182 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384200 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e10a610-e422-4785-89a4-5c1beb50c116-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384217 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384237 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dx4z\" (UniqueName: \"kubernetes.io/projected/1e10a610-e422-4785-89a4-5c1beb50c116-kube-api-access-2dx4z\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.384257 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e10a610-e422-4785-89a4-5c1beb50c116-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.619728 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" path="/var/lib/kubelet/pods/003d6dc1-3777-4d37-a709-55a2d6c11623/volumes" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.621349 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" path="/var/lib/kubelet/pods/58631f04-ebd4-426d-bd77-98b5ff6e00f9/volumes" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.866898 5018 generic.go:334] "Generic (PLEG): container finished" podID="1e10a610-e422-4785-89a4-5c1beb50c116" containerID="31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1" exitCode=137 Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.866980 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerDied","Data":"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1"} Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.867266 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cdc789c6-8thlf" event={"ID":"1e10a610-e422-4785-89a4-5c1beb50c116","Type":"ContainerDied","Data":"05313d0a7ff759c703224c1e949fb44c7979dfebe4f6685fd7450d5381fcd431"} Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.867302 5018 scope.go:117] "RemoveContainer" containerID="cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.867001 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cdc789c6-8thlf" Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.901922 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:56:42 crc kubenswrapper[5018]: I1014 08:56:42.912796 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5cdc789c6-8thlf"] Oct 14 08:56:43 crc kubenswrapper[5018]: I1014 08:56:43.089148 5018 scope.go:117] "RemoveContainer" containerID="31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1" Oct 14 08:56:43 crc kubenswrapper[5018]: I1014 08:56:43.120307 5018 scope.go:117] "RemoveContainer" containerID="cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41" Oct 14 08:56:43 crc kubenswrapper[5018]: E1014 08:56:43.121387 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41\": container with ID starting with cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41 not found: ID does not exist" containerID="cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41" Oct 14 08:56:43 crc kubenswrapper[5018]: I1014 08:56:43.121429 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41"} err="failed to get container status \"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41\": rpc error: code = NotFound desc = could not find container \"cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41\": container with ID starting with cb7185fa9b63993bdf8e2bf134e58e22d221efcf0cbf7e59550fecc39e210d41 not found: ID does not exist" Oct 14 08:56:43 crc kubenswrapper[5018]: I1014 08:56:43.121455 5018 scope.go:117] "RemoveContainer" containerID="31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1" Oct 14 08:56:43 crc kubenswrapper[5018]: E1014 08:56:43.122132 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1\": container with ID starting with 31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1 not found: ID does not exist" containerID="31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1" Oct 14 08:56:43 crc kubenswrapper[5018]: I1014 08:56:43.122154 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1"} err="failed to get container status \"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1\": rpc error: code = NotFound desc = could not find container \"31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1\": container with ID starting with 31f6a1ab37ac17212f5c363e21fbe4062a8160468b4220e91d280c9a457ad5a1 not found: ID does not exist" Oct 14 08:56:44 crc kubenswrapper[5018]: I1014 08:56:44.619453 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" path="/var/lib/kubelet/pods/1e10a610-e422-4785-89a4-5c1beb50c116/volumes" Oct 14 08:56:48 crc kubenswrapper[5018]: I1014 08:56:48.385689 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 08:56:48 crc kubenswrapper[5018]: I1014 08:56:48.452484 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:56:48 crc kubenswrapper[5018]: I1014 08:56:48.453052 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-55c7b9d586-sxknt" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerName="heat-engine" containerID="cri-o://e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" gracePeriod=60 Oct 14 08:56:51 crc kubenswrapper[5018]: E1014 08:56:51.015748 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 08:56:51 crc kubenswrapper[5018]: E1014 08:56:51.018414 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 08:56:51 crc kubenswrapper[5018]: E1014 08:56:51.019568 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 08:56:51 crc kubenswrapper[5018]: E1014 08:56:51.019601 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-55c7b9d586-sxknt" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerName="heat-engine" Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.778753 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.913777 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle\") pod \"8bb51c74-168b-4420-ba0b-b2ce254929d8\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.913856 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9f6g\" (UniqueName: \"kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g\") pod \"8bb51c74-168b-4420-ba0b-b2ce254929d8\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.913896 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom\") pod \"8bb51c74-168b-4420-ba0b-b2ce254929d8\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.914027 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data\") pod \"8bb51c74-168b-4420-ba0b-b2ce254929d8\" (UID: \"8bb51c74-168b-4420-ba0b-b2ce254929d8\") " Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.920338 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g" (OuterVolumeSpecName: "kube-api-access-v9f6g") pod "8bb51c74-168b-4420-ba0b-b2ce254929d8" (UID: "8bb51c74-168b-4420-ba0b-b2ce254929d8"). InnerVolumeSpecName "kube-api-access-v9f6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.928971 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8bb51c74-168b-4420-ba0b-b2ce254929d8" (UID: "8bb51c74-168b-4420-ba0b-b2ce254929d8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.950876 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bb51c74-168b-4420-ba0b-b2ce254929d8" (UID: "8bb51c74-168b-4420-ba0b-b2ce254929d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:00 crc kubenswrapper[5018]: I1014 08:57:00.973500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data" (OuterVolumeSpecName: "config-data") pod "8bb51c74-168b-4420-ba0b-b2ce254929d8" (UID: "8bb51c74-168b-4420-ba0b-b2ce254929d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.015502 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.015551 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.015562 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9f6g\" (UniqueName: \"kubernetes.io/projected/8bb51c74-168b-4420-ba0b-b2ce254929d8-kube-api-access-v9f6g\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.015571 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb51c74-168b-4420-ba0b-b2ce254929d8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.061261 5018 generic.go:334] "Generic (PLEG): container finished" podID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" exitCode=0 Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.061360 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c7b9d586-sxknt" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.061359 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c7b9d586-sxknt" event={"ID":"8bb51c74-168b-4420-ba0b-b2ce254929d8","Type":"ContainerDied","Data":"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7"} Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.061673 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c7b9d586-sxknt" event={"ID":"8bb51c74-168b-4420-ba0b-b2ce254929d8","Type":"ContainerDied","Data":"d443e735fb983688f946c7ced2ed36524c308ab86ccdb6e6d8ca2b519aa265d1"} Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.061696 5018 scope.go:117] "RemoveContainer" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.079750 5018 scope.go:117] "RemoveContainer" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" Oct 14 08:57:01 crc kubenswrapper[5018]: E1014 08:57:01.080100 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7\": container with ID starting with e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7 not found: ID does not exist" containerID="e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.080201 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7"} err="failed to get container status \"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7\": rpc error: code = NotFound desc = could not find container \"e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7\": container with ID starting with e198ca8ad206c139f8c69e061ecef3ad64685981ec1ce2624c400547e7049fa7 not found: ID does not exist" Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.101704 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:57:01 crc kubenswrapper[5018]: I1014 08:57:01.109803 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-55c7b9d586-sxknt"] Oct 14 08:57:02 crc kubenswrapper[5018]: I1014 08:57:02.617510 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" path="/var/lib/kubelet/pods/8bb51c74-168b-4420-ba0b-b2ce254929d8/volumes" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.714553 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4"] Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715546 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715569 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715593 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon-log" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715603 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon-log" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715657 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715672 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715695 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerName="heat-engine" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715708 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerName="heat-engine" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715730 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715740 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715759 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715767 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715776 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715784 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.715802 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.715812 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716100 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716124 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="267cbd65-9ace-4760-8602-3adb5e589748" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716143 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716160 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb51c74-168b-4420-ba0b-b2ce254929d8" containerName="heat-engine" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716177 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716201 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716218 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb4912b-5bce-41ac-a728-aa0bdaea70c3" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716241 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e10a610-e422-4785-89a4-5c1beb50c116" containerName="horizon-log" Oct 14 08:57:04 crc kubenswrapper[5018]: E1014 08:57:04.716504 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716516 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="58631f04-ebd4-426d-bd77-98b5ff6e00f9" containerName="heat-api" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.716799 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="003d6dc1-3777-4d37-a709-55a2d6c11623" containerName="heat-cfnapi" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.718231 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.720613 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.734557 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4"] Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.904646 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzpct\" (UniqueName: \"kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.904782 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:04 crc kubenswrapper[5018]: I1014 08:57:04.905030 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.007913 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.008155 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzpct\" (UniqueName: \"kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.008344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.008695 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.008736 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.033172 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzpct\" (UniqueName: \"kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.061398 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:05 crc kubenswrapper[5018]: I1014 08:57:05.590313 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4"] Oct 14 08:57:06 crc kubenswrapper[5018]: I1014 08:57:06.136187 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerStarted","Data":"ad1a71c193aefbb5238a344f10910ab1cefd470d8053e854f6b9633784773430"} Oct 14 08:57:06 crc kubenswrapper[5018]: I1014 08:57:06.136461 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerStarted","Data":"24372c7aa66d31ad60f9dbfdcdfaf6aa8ab34bd67ceab1bc19b2379ee1dc09f1"} Oct 14 08:57:07 crc kubenswrapper[5018]: I1014 08:57:07.154100 5018 generic.go:334] "Generic (PLEG): container finished" podID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerID="ad1a71c193aefbb5238a344f10910ab1cefd470d8053e854f6b9633784773430" exitCode=0 Oct 14 08:57:07 crc kubenswrapper[5018]: I1014 08:57:07.154540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerDied","Data":"ad1a71c193aefbb5238a344f10910ab1cefd470d8053e854f6b9633784773430"} Oct 14 08:57:09 crc kubenswrapper[5018]: I1014 08:57:09.177497 5018 generic.go:334] "Generic (PLEG): container finished" podID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerID="be1d82695c32b89b2297429a30b9674f1778075e54852baac07b65daa83d34dc" exitCode=0 Oct 14 08:57:09 crc kubenswrapper[5018]: I1014 08:57:09.177588 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerDied","Data":"be1d82695c32b89b2297429a30b9674f1778075e54852baac07b65daa83d34dc"} Oct 14 08:57:10 crc kubenswrapper[5018]: I1014 08:57:10.190890 5018 generic.go:334] "Generic (PLEG): container finished" podID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerID="026181d1c5f9ec0adaa525130cbc7706d37c06870406935a902898cbd2d4a841" exitCode=0 Oct 14 08:57:10 crc kubenswrapper[5018]: I1014 08:57:10.190986 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerDied","Data":"026181d1c5f9ec0adaa525130cbc7706d37c06870406935a902898cbd2d4a841"} Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.621227 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.742135 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzpct\" (UniqueName: \"kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct\") pod \"ed44b30f-2a96-4e81-9516-03fd51a08883\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.742193 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle\") pod \"ed44b30f-2a96-4e81-9516-03fd51a08883\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.742425 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util\") pod \"ed44b30f-2a96-4e81-9516-03fd51a08883\" (UID: \"ed44b30f-2a96-4e81-9516-03fd51a08883\") " Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.745529 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle" (OuterVolumeSpecName: "bundle") pod "ed44b30f-2a96-4e81-9516-03fd51a08883" (UID: "ed44b30f-2a96-4e81-9516-03fd51a08883"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.756262 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct" (OuterVolumeSpecName: "kube-api-access-xzpct") pod "ed44b30f-2a96-4e81-9516-03fd51a08883" (UID: "ed44b30f-2a96-4e81-9516-03fd51a08883"). InnerVolumeSpecName "kube-api-access-xzpct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.758044 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util" (OuterVolumeSpecName: "util") pod "ed44b30f-2a96-4e81-9516-03fd51a08883" (UID: "ed44b30f-2a96-4e81-9516-03fd51a08883"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.846107 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzpct\" (UniqueName: \"kubernetes.io/projected/ed44b30f-2a96-4e81-9516-03fd51a08883-kube-api-access-xzpct\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.846138 5018 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:11 crc kubenswrapper[5018]: I1014 08:57:11.846147 5018 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ed44b30f-2a96-4e81-9516-03fd51a08883-util\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:12 crc kubenswrapper[5018]: I1014 08:57:12.214478 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" event={"ID":"ed44b30f-2a96-4e81-9516-03fd51a08883","Type":"ContainerDied","Data":"24372c7aa66d31ad60f9dbfdcdfaf6aa8ab34bd67ceab1bc19b2379ee1dc09f1"} Oct 14 08:57:12 crc kubenswrapper[5018]: I1014 08:57:12.214537 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24372c7aa66d31ad60f9dbfdcdfaf6aa8ab34bd67ceab1bc19b2379ee1dc09f1" Oct 14 08:57:12 crc kubenswrapper[5018]: I1014 08:57:12.214576 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4" Oct 14 08:57:14 crc kubenswrapper[5018]: I1014 08:57:14.598299 5018 scope.go:117] "RemoveContainer" containerID="d646487382f3976a5b428292b3f200405437c1bf82a45abec7176978973ca91f" Oct 14 08:57:14 crc kubenswrapper[5018]: I1014 08:57:14.645790 5018 scope.go:117] "RemoveContainer" containerID="e7ce72802b638ad8af5e63ac212ad5735398a50811daa5c990e45f24450302f6" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.973351 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7"] Oct 14 08:57:22 crc kubenswrapper[5018]: E1014 08:57:22.974068 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="util" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.974087 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="util" Oct 14 08:57:22 crc kubenswrapper[5018]: E1014 08:57:22.974128 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="pull" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.974135 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="pull" Oct 14 08:57:22 crc kubenswrapper[5018]: E1014 08:57:22.974147 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="extract" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.974153 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="extract" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.974339 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed44b30f-2a96-4e81-9516-03fd51a08883" containerName="extract" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.975148 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.976893 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.977123 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-gzznb" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.979643 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 14 08:57:22 crc kubenswrapper[5018]: I1014 08:57:22.987679 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.034632 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.035841 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.039574 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-nt98r" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.039810 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.047216 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.065145 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.066669 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.071818 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s48m\" (UniqueName: \"kubernetes.io/projected/6922faee-23ec-45f2-9542-5d2e48ccf319-kube-api-access-8s48m\") pod \"obo-prometheus-operator-7c8cf85677-skzn7\" (UID: \"6922faee-23ec-45f2-9542-5d2e48ccf319\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.078075 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.173907 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.173975 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.174021 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s48m\" (UniqueName: \"kubernetes.io/projected/6922faee-23ec-45f2-9542-5d2e48ccf319-kube-api-access-8s48m\") pod \"obo-prometheus-operator-7c8cf85677-skzn7\" (UID: \"6922faee-23ec-45f2-9542-5d2e48ccf319\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.174077 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.174132 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.194729 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-dwfhs"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.196142 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.196432 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s48m\" (UniqueName: \"kubernetes.io/projected/6922faee-23ec-45f2-9542-5d2e48ccf319-kube-api-access-8s48m\") pod \"obo-prometheus-operator-7c8cf85677-skzn7\" (UID: \"6922faee-23ec-45f2-9542-5d2e48ccf319\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.198719 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-fgg2c" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.200324 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.218383 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-dwfhs"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.275762 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsmkl\" (UniqueName: \"kubernetes.io/projected/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-kube-api-access-tsmkl\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.275812 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.275860 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.275931 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.275971 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.276023 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.281203 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.282829 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.283639 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ebe5d8f-9403-41da-8c97-ed60af150539-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-rptmv\" (UID: \"1ebe5d8f-9403-41da-8c97-ed60af150539\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.284013 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de42af88-f42f-4a3a-8a7a-77dec0807d7a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6bf767b955-zst8q\" (UID: \"de42af88-f42f-4a3a-8a7a-77dec0807d7a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.295084 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.317708 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-vs6vr"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.327777 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.348681 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-g5f2t" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.350544 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-vs6vr"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.366728 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.378595 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.378769 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsmkl\" (UniqueName: \"kubernetes.io/projected/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-kube-api-access-tsmkl\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.387605 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.388610 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.401367 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsmkl\" (UniqueName: \"kubernetes.io/projected/41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d-kube-api-access-tsmkl\") pod \"observability-operator-cc5f78dfc-dwfhs\" (UID: \"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d\") " pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.481553 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.481635 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r4pc\" (UniqueName: \"kubernetes.io/projected/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-kube-api-access-6r4pc\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.549875 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.590205 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r4pc\" (UniqueName: \"kubernetes.io/projected/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-kube-api-access-6r4pc\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.590361 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.591163 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.610176 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r4pc\" (UniqueName: \"kubernetes.io/projected/b1d0a181-0c76-4535-a5e0-ffd8bff1383e-kube-api-access-6r4pc\") pod \"perses-operator-54bc95c9fb-vs6vr\" (UID: \"b1d0a181-0c76-4535-a5e0-ffd8bff1383e\") " pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.786784 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.978564 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7"] Oct 14 08:57:23 crc kubenswrapper[5018]: I1014 08:57:23.997185 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv"] Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.013804 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.068962 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jchrc"] Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.081756 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jchrc"] Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.124592 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q"] Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.141483 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-dwfhs"] Oct 14 08:57:24 crc kubenswrapper[5018]: W1014 08:57:24.178758 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41b12c8d_37f0_4fe5_91a2_3dc8585f9f9d.slice/crio-3bcf1f7ab186ae6db8aa3ecec1c9fa07240e38c36824d0b421adf6fe99fea6a3 WatchSource:0}: Error finding container 3bcf1f7ab186ae6db8aa3ecec1c9fa07240e38c36824d0b421adf6fe99fea6a3: Status 404 returned error can't find the container with id 3bcf1f7ab186ae6db8aa3ecec1c9fa07240e38c36824d0b421adf6fe99fea6a3 Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.304324 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-vs6vr"] Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.419396 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" event={"ID":"b1d0a181-0c76-4535-a5e0-ffd8bff1383e","Type":"ContainerStarted","Data":"e31c5a1ad85a1a9956f9b8166796613cf155ab2113fde60cad2027009a0114dd"} Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.420526 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" event={"ID":"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d","Type":"ContainerStarted","Data":"3bcf1f7ab186ae6db8aa3ecec1c9fa07240e38c36824d0b421adf6fe99fea6a3"} Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.421905 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" event={"ID":"1ebe5d8f-9403-41da-8c97-ed60af150539","Type":"ContainerStarted","Data":"6d646f4d07f6c31c8d7259307b4b882acda1def2c4ab9b446aa7e1ced6169548"} Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.422850 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" event={"ID":"6922faee-23ec-45f2-9542-5d2e48ccf319","Type":"ContainerStarted","Data":"7a6a7780e99d974aca5f5baa59502c253c8f026542272ced9f95c6f30b237f58"} Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.424199 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" event={"ID":"de42af88-f42f-4a3a-8a7a-77dec0807d7a","Type":"ContainerStarted","Data":"3095f5b48b673c7cc569d9b5e3476ed1f8a68307ffb360f6e68c1cc4664ef82e"} Oct 14 08:57:24 crc kubenswrapper[5018]: I1014 08:57:24.622791 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22" path="/var/lib/kubelet/pods/e7b36d12-fd6d-4b9d-a9ef-0bc6ee7c1e22/volumes" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.515752 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" event={"ID":"b1d0a181-0c76-4535-a5e0-ffd8bff1383e","Type":"ContainerStarted","Data":"38535efd900c20edef294b1329b575bf1f8997b37536cb9a121092311b0d040d"} Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.516369 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.517704 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" event={"ID":"41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d","Type":"ContainerStarted","Data":"f0f244b817a33307740de9cbf8880d8d64052f25d0133f0e64b6c619140c40d1"} Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.518497 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.520289 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" event={"ID":"1ebe5d8f-9403-41da-8c97-ed60af150539","Type":"ContainerStarted","Data":"2c2d03b6c0e1a0ada3c2765a7c570e80c0b9f6d3d763920ff315341d5f1b2849"} Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.521501 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" event={"ID":"6922faee-23ec-45f2-9542-5d2e48ccf319","Type":"ContainerStarted","Data":"754578c54fadcdbb7a9401f969c7f5a4b3bb9e431caac2a82dc78a2dad291988"} Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.522574 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" event={"ID":"de42af88-f42f-4a3a-8a7a-77dec0807d7a","Type":"ContainerStarted","Data":"4f06df99f403cf91a3ff8e17cceed46af0c255477612365ce2ac301a1eb952d2"} Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.542226 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" podStartSLOduration=2.651898051 podStartE2EDuration="9.542203739s" podCreationTimestamp="2025-10-14 08:57:23 +0000 UTC" firstStartedPulling="2025-10-14 08:57:24.314548568 +0000 UTC m=+7660.898595195" lastFinishedPulling="2025-10-14 08:57:31.204854236 +0000 UTC m=+7667.788900883" observedRunningTime="2025-10-14 08:57:32.534955983 +0000 UTC m=+7669.119002610" watchObservedRunningTime="2025-10-14 08:57:32.542203739 +0000 UTC m=+7669.126250366" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.578225 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.604551 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-rptmv" podStartSLOduration=2.525665353 podStartE2EDuration="9.60450669s" podCreationTimestamp="2025-10-14 08:57:23 +0000 UTC" firstStartedPulling="2025-10-14 08:57:24.029593029 +0000 UTC m=+7660.613639656" lastFinishedPulling="2025-10-14 08:57:31.108434366 +0000 UTC m=+7667.692480993" observedRunningTime="2025-10-14 08:57:32.565371288 +0000 UTC m=+7669.149417915" watchObservedRunningTime="2025-10-14 08:57:32.60450669 +0000 UTC m=+7669.188553347" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.607612 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6bf767b955-zst8q" podStartSLOduration=3.657644545 podStartE2EDuration="10.607602718s" podCreationTimestamp="2025-10-14 08:57:22 +0000 UTC" firstStartedPulling="2025-10-14 08:57:24.158713839 +0000 UTC m=+7660.742760466" lastFinishedPulling="2025-10-14 08:57:31.108672022 +0000 UTC m=+7667.692718639" observedRunningTime="2025-10-14 08:57:32.6020317 +0000 UTC m=+7669.186078337" watchObservedRunningTime="2025-10-14 08:57:32.607602718 +0000 UTC m=+7669.191649345" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.677029 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-skzn7" podStartSLOduration=3.573622356 podStartE2EDuration="10.676999441s" podCreationTimestamp="2025-10-14 08:57:22 +0000 UTC" firstStartedPulling="2025-10-14 08:57:24.013574923 +0000 UTC m=+7660.597621550" lastFinishedPulling="2025-10-14 08:57:31.116951998 +0000 UTC m=+7667.700998635" observedRunningTime="2025-10-14 08:57:32.666453291 +0000 UTC m=+7669.250499938" watchObservedRunningTime="2025-10-14 08:57:32.676999441 +0000 UTC m=+7669.261046078" Oct 14 08:57:32 crc kubenswrapper[5018]: I1014 08:57:32.700770 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-dwfhs" podStartSLOduration=2.614231311 podStartE2EDuration="9.700746836s" podCreationTimestamp="2025-10-14 08:57:23 +0000 UTC" firstStartedPulling="2025-10-14 08:57:24.188549577 +0000 UTC m=+7660.772596204" lastFinishedPulling="2025-10-14 08:57:31.275065102 +0000 UTC m=+7667.859111729" observedRunningTime="2025-10-14 08:57:32.692925683 +0000 UTC m=+7669.276972320" watchObservedRunningTime="2025-10-14 08:57:32.700746836 +0000 UTC m=+7669.284793463" Oct 14 08:57:34 crc kubenswrapper[5018]: I1014 08:57:34.048653 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3690-account-create-zl9fz"] Oct 14 08:57:34 crc kubenswrapper[5018]: I1014 08:57:34.056654 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3690-account-create-zl9fz"] Oct 14 08:57:34 crc kubenswrapper[5018]: I1014 08:57:34.619980 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="998e8c25-bbdf-43b3-a985-3b4c96e25e10" path="/var/lib/kubelet/pods/998e8c25-bbdf-43b3-a985-3b4c96e25e10/volumes" Oct 14 08:57:43 crc kubenswrapper[5018]: I1014 08:57:43.790303 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-vs6vr" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.306944 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.307485 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" containerName="openstackclient" containerID="cri-o://15a85a608abcc9b92abd5e6818da1d44323d70efb24d37f1fdb4f4fd60ac9b4c" gracePeriod=2 Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.314431 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.373228 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:57:46 crc kubenswrapper[5018]: E1014 08:57:46.373939 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" containerName="openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.373970 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" containerName="openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.374329 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" containerName="openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.375518 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.388235 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.391442 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.391490 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fz2p\" (UniqueName: \"kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.391531 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.391617 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.401961 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.493086 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.493134 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fz2p\" (UniqueName: \"kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.493177 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.493294 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.494866 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.529348 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.533209 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.535278 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fz2p\" (UniqueName: \"kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p\") pod \"openstackclient\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.736872 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.738353 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.739967 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.744129 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6cs9j" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.794581 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.817122 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd8tl\" (UniqueName: \"kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl\") pod \"kube-state-metrics-0\" (UID: \"5174f11a-5230-4d98-b393-82a90852f9dd\") " pod="openstack/kube-state-metrics-0" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.918778 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd8tl\" (UniqueName: \"kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl\") pod \"kube-state-metrics-0\" (UID: \"5174f11a-5230-4d98-b393-82a90852f9dd\") " pod="openstack/kube-state-metrics-0" Oct 14 08:57:46 crc kubenswrapper[5018]: I1014 08:57:46.971547 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd8tl\" (UniqueName: \"kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl\") pod \"kube-state-metrics-0\" (UID: \"5174f11a-5230-4d98-b393-82a90852f9dd\") " pod="openstack/kube-state-metrics-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.057051 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.177745 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.180223 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.190170 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-sgc7x" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.190179 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.190348 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.190417 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.198920 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.237945 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.237992 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.238029 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh82b\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.238067 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.238100 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.238125 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.341818 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.341869 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.342589 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh82b\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.342654 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.342677 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.342703 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.343141 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.346859 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.347231 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.351079 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.355059 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.368789 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh82b\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b\") pod \"alertmanager-metric-storage-0\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.521226 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.631799 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:57:47 crc kubenswrapper[5018]: W1014 08:57:47.664090 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddca054de_a18c_46c4_b38e_ebfe315878b7.slice/crio-7f3ddbb8fc2178eec1d9093ec43d6a8e3c36908c46735025bf0cff437a22d271 WatchSource:0}: Error finding container 7f3ddbb8fc2178eec1d9093ec43d6a8e3c36908c46735025bf0cff437a22d271: Status 404 returned error can't find the container with id 7f3ddbb8fc2178eec1d9093ec43d6a8e3c36908c46735025bf0cff437a22d271 Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.750135 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dca054de-a18c-46c4-b38e-ebfe315878b7","Type":"ContainerStarted","Data":"7f3ddbb8fc2178eec1d9093ec43d6a8e3c36908c46735025bf0cff437a22d271"} Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.750608 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:57:47 crc kubenswrapper[5018]: W1014 08:57:47.782505 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5174f11a_5230_4d98_b393_82a90852f9dd.slice/crio-4bd8e9599e84c24b16e0562c61e849c97e51f526de4c5b315646ea1b44fb12c6 WatchSource:0}: Error finding container 4bd8e9599e84c24b16e0562c61e849c97e51f526de4c5b315646ea1b44fb12c6: Status 404 returned error can't find the container with id 4bd8e9599e84c24b16e0562c61e849c97e51f526de4c5b315646ea1b44fb12c6 Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.911826 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.928714 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.931172 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.933950 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.934125 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.934365 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.934475 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.942760 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hms5c" Oct 14 08:57:47 crc kubenswrapper[5018]: I1014 08:57:47.943164 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.069348 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.069663 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.069774 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6ncw\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.069862 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.069946 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.070033 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.070112 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.070187 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171655 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171719 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171740 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171764 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171855 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171879 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171915 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6ncw\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.171942 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.172698 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.179851 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.187455 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.198500 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.216827 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.225453 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.226868 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.226906 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/137b310b4a23998875714c87b97b23efed453f709ae4dfe74d4d0f0ea21ff80c/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.252503 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6ncw\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.292951 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.320807 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.380573 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 08:57:48 crc kubenswrapper[5018]: W1014 08:57:48.439368 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f24304f_0a66_4c0f_b8cb_3341f9fb926d.slice/crio-4dfa3df785869f037ff7b347a4d536e19df442cb05356e13eeef22705f84f0bd WatchSource:0}: Error finding container 4dfa3df785869f037ff7b347a4d536e19df442cb05356e13eeef22705f84f0bd: Status 404 returned error can't find the container with id 4dfa3df785869f037ff7b347a4d536e19df442cb05356e13eeef22705f84f0bd Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.784723 5018 generic.go:334] "Generic (PLEG): container finished" podID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" containerID="15a85a608abcc9b92abd5e6818da1d44323d70efb24d37f1fdb4f4fd60ac9b4c" exitCode=137 Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.785089 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac0dd6ee14ece45ca7cd80206f8a2f85541bd56a49c59c32eb64b55847daba88" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.808435 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5174f11a-5230-4d98-b393-82a90852f9dd","Type":"ContainerStarted","Data":"25666f73a516304a499c435a4de6d90f10c044eb527d649f6a7c5c95162f6ffc"} Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.808474 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5174f11a-5230-4d98-b393-82a90852f9dd","Type":"ContainerStarted","Data":"4bd8e9599e84c24b16e0562c61e849c97e51f526de4c5b315646ea1b44fb12c6"} Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.809480 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.814473 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dca054de-a18c-46c4-b38e-ebfe315878b7","Type":"ContainerStarted","Data":"224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b"} Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.815925 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerStarted","Data":"4dfa3df785869f037ff7b347a4d536e19df442cb05356e13eeef22705f84f0bd"} Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.835171 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.337312026 podStartE2EDuration="2.835150185s" podCreationTimestamp="2025-10-14 08:57:46 +0000 UTC" firstStartedPulling="2025-10-14 08:57:47.798018167 +0000 UTC m=+7684.382064794" lastFinishedPulling="2025-10-14 08:57:48.295856326 +0000 UTC m=+7684.879902953" observedRunningTime="2025-10-14 08:57:48.820501039 +0000 UTC m=+7685.404547666" watchObservedRunningTime="2025-10-14 08:57:48.835150185 +0000 UTC m=+7685.419196812" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.849673 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.849658687 podStartE2EDuration="2.849658687s" podCreationTimestamp="2025-10-14 08:57:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:48.838959273 +0000 UTC m=+7685.423005900" watchObservedRunningTime="2025-10-14 08:57:48.849658687 +0000 UTC m=+7685.433705314" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.869153 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:57:48 crc kubenswrapper[5018]: I1014 08:57:48.916311 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.032981 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret\") pod \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.033052 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74k5s\" (UniqueName: \"kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s\") pod \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.033186 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config\") pod \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.033234 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle\") pod \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\" (UID: \"20c5d68f-66eb-4918-85fc-ce1c18960eb7\") " Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.044805 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s" (OuterVolumeSpecName: "kube-api-access-74k5s") pod "20c5d68f-66eb-4918-85fc-ce1c18960eb7" (UID: "20c5d68f-66eb-4918-85fc-ce1c18960eb7"). InnerVolumeSpecName "kube-api-access-74k5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.069360 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "20c5d68f-66eb-4918-85fc-ce1c18960eb7" (UID: "20c5d68f-66eb-4918-85fc-ce1c18960eb7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.069807 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20c5d68f-66eb-4918-85fc-ce1c18960eb7" (UID: "20c5d68f-66eb-4918-85fc-ce1c18960eb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.099335 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "20c5d68f-66eb-4918-85fc-ce1c18960eb7" (UID: "20c5d68f-66eb-4918-85fc-ce1c18960eb7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.135430 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.135467 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.135480 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74k5s\" (UniqueName: \"kubernetes.io/projected/20c5d68f-66eb-4918-85fc-ce1c18960eb7-kube-api-access-74k5s\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.135492 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20c5d68f-66eb-4918-85fc-ce1c18960eb7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.827136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerStarted","Data":"ca51e2f117543c92746d04095d0a9215c66e58914a626b7e980c656ec51cd532"} Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.827167 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:57:49 crc kubenswrapper[5018]: I1014 08:57:49.870203 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" Oct 14 08:57:50 crc kubenswrapper[5018]: I1014 08:57:50.616485 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c5d68f-66eb-4918-85fc-ce1c18960eb7" path="/var/lib/kubelet/pods/20c5d68f-66eb-4918-85fc-ce1c18960eb7/volumes" Oct 14 08:57:55 crc kubenswrapper[5018]: I1014 08:57:55.896279 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerStarted","Data":"93a6064e24381311dfd0a20d1a614392026b9ff3231be0c6e8ac754eeaa62f1c"} Oct 14 08:57:56 crc kubenswrapper[5018]: I1014 08:57:56.917568 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerStarted","Data":"6bb5e9a7103b50932416060b76bb3e23b591840dff1bed95e8bdf196926d9470"} Oct 14 08:57:57 crc kubenswrapper[5018]: I1014 08:57:57.062371 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 08:58:02 crc kubenswrapper[5018]: I1014 08:58:02.025782 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vwvpv"] Oct 14 08:58:02 crc kubenswrapper[5018]: I1014 08:58:02.032935 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vwvpv"] Oct 14 08:58:02 crc kubenswrapper[5018]: I1014 08:58:02.463746 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:58:02 crc kubenswrapper[5018]: I1014 08:58:02.463830 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:58:02 crc kubenswrapper[5018]: I1014 08:58:02.628911 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d26e3c-22a7-4fa5-8f89-4718f312fe4a" path="/var/lib/kubelet/pods/f7d26e3c-22a7-4fa5-8f89-4718f312fe4a/volumes" Oct 14 08:58:06 crc kubenswrapper[5018]: I1014 08:58:06.038492 5018 generic.go:334] "Generic (PLEG): container finished" podID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerID="93a6064e24381311dfd0a20d1a614392026b9ff3231be0c6e8ac754eeaa62f1c" exitCode=0 Oct 14 08:58:06 crc kubenswrapper[5018]: I1014 08:58:06.038770 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerDied","Data":"93a6064e24381311dfd0a20d1a614392026b9ff3231be0c6e8ac754eeaa62f1c"} Oct 14 08:58:08 crc kubenswrapper[5018]: I1014 08:58:08.067170 5018 generic.go:334] "Generic (PLEG): container finished" podID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerID="6bb5e9a7103b50932416060b76bb3e23b591840dff1bed95e8bdf196926d9470" exitCode=0 Oct 14 08:58:08 crc kubenswrapper[5018]: I1014 08:58:08.067309 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerDied","Data":"6bb5e9a7103b50932416060b76bb3e23b591840dff1bed95e8bdf196926d9470"} Oct 14 08:58:09 crc kubenswrapper[5018]: I1014 08:58:09.079180 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerStarted","Data":"ae3477a4359c14611282e48f047572b99c17263c4ab7fa2f6f48060c5cbbd512"} Oct 14 08:58:13 crc kubenswrapper[5018]: I1014 08:58:13.118583 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerStarted","Data":"f8e67da18c07a0242bb09dc5c3e1ef768c85087115a502b92e1ccb4e2c0ad2a7"} Oct 14 08:58:13 crc kubenswrapper[5018]: I1014 08:58:13.119171 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 14 08:58:13 crc kubenswrapper[5018]: I1014 08:58:13.120778 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 14 08:58:13 crc kubenswrapper[5018]: I1014 08:58:13.150068 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.027144316 podStartE2EDuration="26.150049954s" podCreationTimestamp="2025-10-14 08:57:47 +0000 UTC" firstStartedPulling="2025-10-14 08:57:48.451813669 +0000 UTC m=+7685.035860296" lastFinishedPulling="2025-10-14 08:58:08.574719307 +0000 UTC m=+7705.158765934" observedRunningTime="2025-10-14 08:58:13.142124189 +0000 UTC m=+7709.726170836" watchObservedRunningTime="2025-10-14 08:58:13.150049954 +0000 UTC m=+7709.734096581" Oct 14 08:58:14 crc kubenswrapper[5018]: I1014 08:58:14.869993 5018 scope.go:117] "RemoveContainer" containerID="17469fb7cbaaba599763c7ea2a81ff458a929f8ed6bb8547cf34cdf69db914eb" Oct 14 08:58:14 crc kubenswrapper[5018]: I1014 08:58:14.897305 5018 scope.go:117] "RemoveContainer" containerID="15a85a608abcc9b92abd5e6818da1d44323d70efb24d37f1fdb4f4fd60ac9b4c" Oct 14 08:58:14 crc kubenswrapper[5018]: I1014 08:58:14.950281 5018 scope.go:117] "RemoveContainer" containerID="5ead85004aed5d89f1f3d3206f70176e9cee5c2791b72807af02242104b80b7b" Oct 14 08:58:14 crc kubenswrapper[5018]: I1014 08:58:14.990110 5018 scope.go:117] "RemoveContainer" containerID="555d0de4d32cd68c3cbd905d73c9af6bf2510788a6b940bc0ecdacda4f168243" Oct 14 08:58:15 crc kubenswrapper[5018]: I1014 08:58:15.147799 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerStarted","Data":"2eb5fccb1608174d2715db98c2cd51f13ebfec0f875aad1c1dcff0a31a9dd5fb"} Oct 14 08:58:19 crc kubenswrapper[5018]: I1014 08:58:19.207796 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerStarted","Data":"f61d014c5d3ffeb98e8e0f69ba6c15fd4691755a9b0da69001c812220af7f43d"} Oct 14 08:58:23 crc kubenswrapper[5018]: I1014 08:58:23.273458 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerStarted","Data":"0cd86dd10622b98867fef1db94a8a0e61245d00d830d45fe693a2899dd7e229d"} Oct 14 08:58:23 crc kubenswrapper[5018]: I1014 08:58:23.321943 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:23 crc kubenswrapper[5018]: I1014 08:58:23.334404 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.853358997 podStartE2EDuration="37.334381492s" podCreationTimestamp="2025-10-14 08:57:46 +0000 UTC" firstStartedPulling="2025-10-14 08:57:48.920315726 +0000 UTC m=+7685.504362353" lastFinishedPulling="2025-10-14 08:58:22.401338211 +0000 UTC m=+7718.985384848" observedRunningTime="2025-10-14 08:58:23.320028394 +0000 UTC m=+7719.904075031" watchObservedRunningTime="2025-10-14 08:58:23.334381492 +0000 UTC m=+7719.918428129" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.611412 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.616286 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.619152 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.619151 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.622564 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.713912 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.714252 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.714449 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.715457 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vfn8\" (UniqueName: \"kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.715639 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.715797 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.715952 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817238 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vfn8\" (UniqueName: \"kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817297 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817328 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817371 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817398 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817420 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.817495 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.818924 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.821924 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.823559 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.823746 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.823840 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.826232 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.839322 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vfn8\" (UniqueName: \"kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8\") pod \"ceilometer-0\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " pod="openstack/ceilometer-0" Oct 14 08:58:27 crc kubenswrapper[5018]: I1014 08:58:27.950801 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:58:28 crc kubenswrapper[5018]: I1014 08:58:28.481148 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:58:29 crc kubenswrapper[5018]: I1014 08:58:29.338101 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerStarted","Data":"94f10b267613730106b78ac8c9f439ff91aa3c3f6b7026086b5592c2067d2b95"} Oct 14 08:58:32 crc kubenswrapper[5018]: I1014 08:58:32.462855 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:58:32 crc kubenswrapper[5018]: I1014 08:58:32.463323 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:58:33 crc kubenswrapper[5018]: I1014 08:58:33.335856 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:33 crc kubenswrapper[5018]: I1014 08:58:33.346800 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:33 crc kubenswrapper[5018]: I1014 08:58:33.419243 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerStarted","Data":"ffbcb2b2210cf93d25de2e835c74755e1cff96ca38e2c6ac4b7c6dceea54332d"} Oct 14 08:58:33 crc kubenswrapper[5018]: I1014 08:58:33.421297 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.914902 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.915719 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" containerName="openstackclient" containerID="cri-o://224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b" gracePeriod=2 Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.930643 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.949977 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:58:34 crc kubenswrapper[5018]: E1014 08:58:34.950399 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" containerName="openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.950418 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" containerName="openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.950639 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" containerName="openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.952345 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.958581 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dca054de-a18c-46c4-b38e-ebfe315878b7" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.968214 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.984810 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcjht\" (UniqueName: \"kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.985079 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.985115 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:34 crc kubenswrapper[5018]: I1014 08:58:34.985189 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.087373 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.087517 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcjht\" (UniqueName: \"kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.087590 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.087609 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.088463 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.111730 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.112264 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.122313 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcjht\" (UniqueName: \"kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht\") pod \"openstackclient\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.276096 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.482913 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerStarted","Data":"71b14d6647d11fec0ae387faf95137d74438bddc22ff1464760276078402a41a"} Oct 14 08:58:35 crc kubenswrapper[5018]: I1014 08:58:35.918285 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:58:35 crc kubenswrapper[5018]: W1014 08:58:35.943464 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode216b65d_5bc3_453c_aa74_b9792e47575e.slice/crio-de5f56fc9267910261031c4b57dc735c2f81b43ca818b6e906f2c6c914024941 WatchSource:0}: Error finding container de5f56fc9267910261031c4b57dc735c2f81b43ca818b6e906f2c6c914024941: Status 404 returned error can't find the container with id de5f56fc9267910261031c4b57dc735c2f81b43ca818b6e906f2c6c914024941 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.236460 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.238963 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="prometheus" containerID="cri-o://2eb5fccb1608174d2715db98c2cd51f13ebfec0f875aad1c1dcff0a31a9dd5fb" gracePeriod=600 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.239101 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="thanos-sidecar" containerID="cri-o://0cd86dd10622b98867fef1db94a8a0e61245d00d830d45fe693a2899dd7e229d" gracePeriod=600 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.239144 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="config-reloader" containerID="cri-o://f61d014c5d3ffeb98e8e0f69ba6c15fd4691755a9b0da69001c812220af7f43d" gracePeriod=600 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.496844 5018 generic.go:334] "Generic (PLEG): container finished" podID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerID="0cd86dd10622b98867fef1db94a8a0e61245d00d830d45fe693a2899dd7e229d" exitCode=0 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.496999 5018 generic.go:334] "Generic (PLEG): container finished" podID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerID="f61d014c5d3ffeb98e8e0f69ba6c15fd4691755a9b0da69001c812220af7f43d" exitCode=0 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.497008 5018 generic.go:334] "Generic (PLEG): container finished" podID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerID="2eb5fccb1608174d2715db98c2cd51f13ebfec0f875aad1c1dcff0a31a9dd5fb" exitCode=0 Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.496923 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerDied","Data":"0cd86dd10622b98867fef1db94a8a0e61245d00d830d45fe693a2899dd7e229d"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.497076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerDied","Data":"f61d014c5d3ffeb98e8e0f69ba6c15fd4691755a9b0da69001c812220af7f43d"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.497092 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerDied","Data":"2eb5fccb1608174d2715db98c2cd51f13ebfec0f875aad1c1dcff0a31a9dd5fb"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.498384 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e216b65d-5bc3-453c-aa74-b9792e47575e","Type":"ContainerStarted","Data":"ae4fa989c0d49cc43c46cfbb1bf94f22382f8020c2ea5b17401082ece4256593"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.498420 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e216b65d-5bc3-453c-aa74-b9792e47575e","Type":"ContainerStarted","Data":"de5f56fc9267910261031c4b57dc735c2f81b43ca818b6e906f2c6c914024941"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.500320 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerStarted","Data":"b8c91769273de4970f74a2edfeccc49f78d58484cf0ceb4beb6cff90550e93be"} Oct 14 08:58:36 crc kubenswrapper[5018]: I1014 08:58:36.513489 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.51347279 podStartE2EDuration="2.51347279s" podCreationTimestamp="2025-10-14 08:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:58:36.511588277 +0000 UTC m=+7733.095634904" watchObservedRunningTime="2025-10-14 08:58:36.51347279 +0000 UTC m=+7733.097519417" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.294637 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.315125 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434504 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret\") pod \"dca054de-a18c-46c4-b38e-ebfe315878b7\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434664 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle\") pod \"dca054de-a18c-46c4-b38e-ebfe315878b7\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434704 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434729 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fz2p\" (UniqueName: \"kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p\") pod \"dca054de-a18c-46c4-b38e-ebfe315878b7\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434847 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434889 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434947 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config\") pod \"dca054de-a18c-46c4-b38e-ebfe315878b7\" (UID: \"dca054de-a18c-46c4-b38e-ebfe315878b7\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.434972 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6ncw\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.435105 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.435135 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.435181 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.435222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0\") pod \"b7147267-f721-43b5-8048-63bb6c5b1a6f\" (UID: \"b7147267-f721-43b5-8048-63bb6c5b1a6f\") " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.437534 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.439831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p" (OuterVolumeSpecName: "kube-api-access-7fz2p") pod "dca054de-a18c-46c4-b38e-ebfe315878b7" (UID: "dca054de-a18c-46c4-b38e-ebfe315878b7"). InnerVolumeSpecName "kube-api-access-7fz2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.442990 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config" (OuterVolumeSpecName: "config") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.444593 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.444684 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw" (OuterVolumeSpecName: "kube-api-access-k6ncw") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "kube-api-access-k6ncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.450715 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out" (OuterVolumeSpecName: "config-out") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.454209 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.487451 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.503844 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dca054de-a18c-46c4-b38e-ebfe315878b7" (UID: "dca054de-a18c-46c4-b38e-ebfe315878b7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.521534 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dca054de-a18c-46c4-b38e-ebfe315878b7" (UID: "dca054de-a18c-46c4-b38e-ebfe315878b7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.525781 5018 generic.go:334] "Generic (PLEG): container finished" podID="dca054de-a18c-46c4-b38e-ebfe315878b7" containerID="224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b" exitCode=137 Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.525889 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.526161 5018 scope.go:117] "RemoveContainer" containerID="224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.537704 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config" (OuterVolumeSpecName: "web-config") pod "b7147267-f721-43b5-8048-63bb6c5b1a6f" (UID: "b7147267-f721-43b5-8048-63bb6c5b1a6f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.538957 5018 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.538977 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fz2p\" (UniqueName: \"kubernetes.io/projected/dca054de-a18c-46c4-b38e-ebfe315878b7-kube-api-access-7fz2p\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.538997 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") on node \"crc\" " Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539008 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539020 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539029 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6ncw\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-kube-api-access-k6ncw\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539037 5018 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7147267-f721-43b5-8048-63bb6c5b1a6f-config-out\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539045 5018 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7147267-f721-43b5-8048-63bb6c5b1a6f-web-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539057 5018 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7147267-f721-43b5-8048-63bb6c5b1a6f-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539067 5018 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7147267-f721-43b5-8048-63bb6c5b1a6f-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.539075 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.556467 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.559101 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7147267-f721-43b5-8048-63bb6c5b1a6f","Type":"ContainerDied","Data":"ca51e2f117543c92746d04095d0a9215c66e58914a626b7e980c656ec51cd532"} Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.565749 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dca054de-a18c-46c4-b38e-ebfe315878b7" (UID: "dca054de-a18c-46c4-b38e-ebfe315878b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.590104 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.590327 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df") on node "crc" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.624131 5018 scope.go:117] "RemoveContainer" containerID="224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b" Oct 14 08:58:37 crc kubenswrapper[5018]: E1014 08:58:37.624478 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b\": container with ID starting with 224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b not found: ID does not exist" containerID="224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.624517 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b"} err="failed to get container status \"224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b\": rpc error: code = NotFound desc = could not find container \"224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b\": container with ID starting with 224b4f75e45b826c920441ff9ab50895b1cecf3df06fb7640912137e3ac3fa9b not found: ID does not exist" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.624542 5018 scope.go:117] "RemoveContainer" containerID="0cd86dd10622b98867fef1db94a8a0e61245d00d830d45fe693a2899dd7e229d" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.641295 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca054de-a18c-46c4-b38e-ebfe315878b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.641340 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.653641 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.655171 5018 scope.go:117] "RemoveContainer" containerID="f61d014c5d3ffeb98e8e0f69ba6c15fd4691755a9b0da69001c812220af7f43d" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.667693 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.673760 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:37 crc kubenswrapper[5018]: E1014 08:58:37.674133 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="thanos-sidecar" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674144 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="thanos-sidecar" Oct 14 08:58:37 crc kubenswrapper[5018]: E1014 08:58:37.674155 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="prometheus" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674161 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="prometheus" Oct 14 08:58:37 crc kubenswrapper[5018]: E1014 08:58:37.674186 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="config-reloader" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674192 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="config-reloader" Oct 14 08:58:37 crc kubenswrapper[5018]: E1014 08:58:37.674211 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="init-config-reloader" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674234 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="init-config-reloader" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674425 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="config-reloader" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674441 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="thanos-sidecar" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.674455 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" containerName="prometheus" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.676297 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.681002 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.681192 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.681761 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.681925 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.682527 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.682719 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hms5c" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.688676 5018 scope.go:117] "RemoveContainer" containerID="2eb5fccb1608174d2715db98c2cd51f13ebfec0f875aad1c1dcff0a31a9dd5fb" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.689773 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.710572 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.738807 5018 scope.go:117] "RemoveContainer" containerID="6bb5e9a7103b50932416060b76bb3e23b591840dff1bed95e8bdf196926d9470" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844754 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ctxl\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844800 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844837 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844922 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844958 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.844985 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.845003 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.845027 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.845203 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.845353 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.845447 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.856585 5018 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dca054de-a18c-46c4-b38e-ebfe315878b7" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947088 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ctxl\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947149 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947184 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947228 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947270 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947301 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947322 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947353 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947393 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947425 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.947462 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.948232 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.950780 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.951229 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.951932 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.951960 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.952749 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.952950 5018 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.952992 5018 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/137b310b4a23998875714c87b97b23efed453f709ae4dfe74d4d0f0ea21ff80c/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.953222 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.953855 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.954387 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.969479 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ctxl\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:37 crc kubenswrapper[5018]: I1014 08:58:37.996671 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"prometheus-metric-storage-0\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.008169 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.495971 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.572793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerStarted","Data":"917dcfbfd44e3b51a2c06de6a38b735b9a862ee3bfad2604b0b2b3961b0e967b"} Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.577793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerStarted","Data":"3b33a495951843dca3306d1f5afb9c903071ebc4d8c7ab9493eea6fb248bc2b5"} Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.579061 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.619977 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.739412195 podStartE2EDuration="11.619958304s" podCreationTimestamp="2025-10-14 08:58:27 +0000 UTC" firstStartedPulling="2025-10-14 08:58:28.504704731 +0000 UTC m=+7725.088751368" lastFinishedPulling="2025-10-14 08:58:37.38525085 +0000 UTC m=+7733.969297477" observedRunningTime="2025-10-14 08:58:38.603484266 +0000 UTC m=+7735.187530893" watchObservedRunningTime="2025-10-14 08:58:38.619958304 +0000 UTC m=+7735.204004951" Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.620595 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7147267-f721-43b5-8048-63bb6c5b1a6f" path="/var/lib/kubelet/pods/b7147267-f721-43b5-8048-63bb6c5b1a6f/volumes" Oct 14 08:58:38 crc kubenswrapper[5018]: I1014 08:58:38.621419 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca054de-a18c-46c4-b38e-ebfe315878b7" path="/var/lib/kubelet/pods/dca054de-a18c-46c4-b38e-ebfe315878b7/volumes" Oct 14 08:58:42 crc kubenswrapper[5018]: I1014 08:58:42.941280 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-977pq"] Oct 14 08:58:42 crc kubenswrapper[5018]: I1014 08:58:42.943647 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-977pq" Oct 14 08:58:42 crc kubenswrapper[5018]: I1014 08:58:42.952020 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-977pq"] Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.059974 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbmhm\" (UniqueName: \"kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm\") pod \"aodh-db-create-977pq\" (UID: \"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be\") " pod="openstack/aodh-db-create-977pq" Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.161449 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbmhm\" (UniqueName: \"kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm\") pod \"aodh-db-create-977pq\" (UID: \"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be\") " pod="openstack/aodh-db-create-977pq" Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.181141 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbmhm\" (UniqueName: \"kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm\") pod \"aodh-db-create-977pq\" (UID: \"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be\") " pod="openstack/aodh-db-create-977pq" Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.264051 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-977pq" Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.618077 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerStarted","Data":"00899a309e02428a711c0dbf8cf3c13507e7bba8ba46443809cca45c5e90dcc3"} Oct 14 08:58:43 crc kubenswrapper[5018]: I1014 08:58:43.724174 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-977pq"] Oct 14 08:58:44 crc kubenswrapper[5018]: I1014 08:58:44.646366 5018 generic.go:334] "Generic (PLEG): container finished" podID="2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" containerID="d9c7dac13e2dea7e6557db75768e7930258f2cf88772355a4297edfc995e3ef6" exitCode=0 Oct 14 08:58:44 crc kubenswrapper[5018]: I1014 08:58:44.646845 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-977pq" event={"ID":"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be","Type":"ContainerDied","Data":"d9c7dac13e2dea7e6557db75768e7930258f2cf88772355a4297edfc995e3ef6"} Oct 14 08:58:44 crc kubenswrapper[5018]: I1014 08:58:44.646885 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-977pq" event={"ID":"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be","Type":"ContainerStarted","Data":"17f10898a1ad9e69d02b5a95b2041a24740592ee7a537ffdb4f395cb5411a9e5"} Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.148458 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-977pq" Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.238202 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbmhm\" (UniqueName: \"kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm\") pod \"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be\" (UID: \"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be\") " Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.248939 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm" (OuterVolumeSpecName: "kube-api-access-vbmhm") pod "2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" (UID: "2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be"). InnerVolumeSpecName "kube-api-access-vbmhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.340800 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbmhm\" (UniqueName: \"kubernetes.io/projected/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be-kube-api-access-vbmhm\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.674532 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-977pq" event={"ID":"2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be","Type":"ContainerDied","Data":"17f10898a1ad9e69d02b5a95b2041a24740592ee7a537ffdb4f395cb5411a9e5"} Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.674574 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f10898a1ad9e69d02b5a95b2041a24740592ee7a537ffdb4f395cb5411a9e5" Oct 14 08:58:46 crc kubenswrapper[5018]: I1014 08:58:46.674659 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-977pq" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.048198 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-033c-account-create-ptmqh"] Oct 14 08:58:53 crc kubenswrapper[5018]: E1014 08:58:53.057395 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" containerName="mariadb-database-create" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.057426 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" containerName="mariadb-database-create" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.057839 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" containerName="mariadb-database-create" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.059347 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.059770 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-033c-account-create-ptmqh"] Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.061900 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.238554 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc8kn\" (UniqueName: \"kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn\") pod \"aodh-033c-account-create-ptmqh\" (UID: \"73aaf538-652f-43c9-90f2-ad5a844733dc\") " pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.340208 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc8kn\" (UniqueName: \"kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn\") pod \"aodh-033c-account-create-ptmqh\" (UID: \"73aaf538-652f-43c9-90f2-ad5a844733dc\") " pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.377875 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc8kn\" (UniqueName: \"kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn\") pod \"aodh-033c-account-create-ptmqh\" (UID: \"73aaf538-652f-43c9-90f2-ad5a844733dc\") " pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.428136 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.783078 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerID="00899a309e02428a711c0dbf8cf3c13507e7bba8ba46443809cca45c5e90dcc3" exitCode=0 Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.783231 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerDied","Data":"00899a309e02428a711c0dbf8cf3c13507e7bba8ba46443809cca45c5e90dcc3"} Oct 14 08:58:53 crc kubenswrapper[5018]: I1014 08:58:53.914025 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-033c-account-create-ptmqh"] Oct 14 08:58:53 crc kubenswrapper[5018]: W1014 08:58:53.916424 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73aaf538_652f_43c9_90f2_ad5a844733dc.slice/crio-8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24 WatchSource:0}: Error finding container 8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24: Status 404 returned error can't find the container with id 8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24 Oct 14 08:58:54 crc kubenswrapper[5018]: I1014 08:58:54.800513 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerStarted","Data":"5475b3ac1ffb8d1aa80f71e6a95e7eed4b52f74bd503013962bda9fef820ce2f"} Oct 14 08:58:54 crc kubenswrapper[5018]: I1014 08:58:54.804321 5018 generic.go:334] "Generic (PLEG): container finished" podID="73aaf538-652f-43c9-90f2-ad5a844733dc" containerID="2e77272e500e6600d36b4d5079bc8d3c5187606d43f141ea94c262c6476cace8" exitCode=0 Oct 14 08:58:54 crc kubenswrapper[5018]: I1014 08:58:54.804367 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-033c-account-create-ptmqh" event={"ID":"73aaf538-652f-43c9-90f2-ad5a844733dc","Type":"ContainerDied","Data":"2e77272e500e6600d36b4d5079bc8d3c5187606d43f141ea94c262c6476cace8"} Oct 14 08:58:54 crc kubenswrapper[5018]: I1014 08:58:54.804410 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-033c-account-create-ptmqh" event={"ID":"73aaf538-652f-43c9-90f2-ad5a844733dc","Type":"ContainerStarted","Data":"8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24"} Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.345294 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.519091 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc8kn\" (UniqueName: \"kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn\") pod \"73aaf538-652f-43c9-90f2-ad5a844733dc\" (UID: \"73aaf538-652f-43c9-90f2-ad5a844733dc\") " Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.529466 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn" (OuterVolumeSpecName: "kube-api-access-lc8kn") pod "73aaf538-652f-43c9-90f2-ad5a844733dc" (UID: "73aaf538-652f-43c9-90f2-ad5a844733dc"). InnerVolumeSpecName "kube-api-access-lc8kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.623224 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc8kn\" (UniqueName: \"kubernetes.io/projected/73aaf538-652f-43c9-90f2-ad5a844733dc-kube-api-access-lc8kn\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.841272 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-033c-account-create-ptmqh" event={"ID":"73aaf538-652f-43c9-90f2-ad5a844733dc","Type":"ContainerDied","Data":"8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24"} Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.841301 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-033c-account-create-ptmqh" Oct 14 08:58:56 crc kubenswrapper[5018]: I1014 08:58:56.841312 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2edb62a3e9aef8b6527227053fe347a683729698a52577a43abeeb06ba3c24" Oct 14 08:58:57 crc kubenswrapper[5018]: I1014 08:58:57.959928 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.492733 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-s48hc"] Oct 14 08:58:58 crc kubenswrapper[5018]: E1014 08:58:58.493268 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73aaf538-652f-43c9-90f2-ad5a844733dc" containerName="mariadb-account-create" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.493283 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="73aaf538-652f-43c9-90f2-ad5a844733dc" containerName="mariadb-account-create" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.493506 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="73aaf538-652f-43c9-90f2-ad5a844733dc" containerName="mariadb-account-create" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.494220 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.497410 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-ps45p" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.497746 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.497956 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.511298 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-s48hc"] Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.669608 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh77v\" (UniqueName: \"kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.669911 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.670032 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.670161 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.771870 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh77v\" (UniqueName: \"kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.771978 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.772028 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.772076 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.778888 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.779404 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.792123 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.793372 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh77v\" (UniqueName: \"kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v\") pod \"aodh-db-sync-s48hc\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.832687 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-s48hc" Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.871366 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerStarted","Data":"77311f9821ca7a858e16455376cf6227d93c7f441b52752f6083c5cada76a306"} Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.871411 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerStarted","Data":"cf8f965d2cb92a7c4505cb3c64420b7047da837b854170fa012a0781465f8cb5"} Oct 14 08:58:58 crc kubenswrapper[5018]: I1014 08:58:58.913178 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.913156635 podStartE2EDuration="21.913156635s" podCreationTimestamp="2025-10-14 08:58:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:58:58.899027463 +0000 UTC m=+7755.483074090" watchObservedRunningTime="2025-10-14 08:58:58.913156635 +0000 UTC m=+7755.497203262" Oct 14 08:58:59 crc kubenswrapper[5018]: I1014 08:58:59.361249 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-s48hc"] Oct 14 08:58:59 crc kubenswrapper[5018]: W1014 08:58:59.364168 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod208ccbe0_ffd5_478f_aca6_8af7d38f9b33.slice/crio-fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807 WatchSource:0}: Error finding container fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807: Status 404 returned error can't find the container with id fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807 Oct 14 08:58:59 crc kubenswrapper[5018]: I1014 08:58:59.884734 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-s48hc" event={"ID":"208ccbe0-ffd5-478f-aca6-8af7d38f9b33","Type":"ContainerStarted","Data":"fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807"} Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.462965 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.463535 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.463586 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.464263 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.464322 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" gracePeriod=600 Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.736886 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.737121 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="5174f11a-5230-4d98-b393-82a90852f9dd" containerName="kube-state-metrics" containerID="cri-o://25666f73a516304a499c435a4de6d90f10c044eb527d649f6a7c5c95162f6ffc" gracePeriod=30 Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.945491 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" exitCode=0 Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.945562 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef"} Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.945596 5018 scope.go:117] "RemoveContainer" containerID="6cdba1e6a453035d3ec0be5c58cd89e1a3a0d19f7bf7ee77a2be84e426fe923e" Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.950343 5018 generic.go:334] "Generic (PLEG): container finished" podID="5174f11a-5230-4d98-b393-82a90852f9dd" containerID="25666f73a516304a499c435a4de6d90f10c044eb527d649f6a7c5c95162f6ffc" exitCode=2 Oct 14 08:59:02 crc kubenswrapper[5018]: I1014 08:59:02.950368 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5174f11a-5230-4d98-b393-82a90852f9dd","Type":"ContainerDied","Data":"25666f73a516304a499c435a4de6d90f10c044eb527d649f6a7c5c95162f6ffc"} Oct 14 08:59:03 crc kubenswrapper[5018]: I1014 08:59:03.008785 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.426052 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.426653 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-central-agent" containerID="cri-o://ffbcb2b2210cf93d25de2e835c74755e1cff96ca38e2c6ac4b7c6dceea54332d" gracePeriod=30 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.426730 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="sg-core" containerID="cri-o://b8c91769273de4970f74a2edfeccc49f78d58484cf0ceb4beb6cff90550e93be" gracePeriod=30 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.426768 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-notification-agent" containerID="cri-o://71b14d6647d11fec0ae387faf95137d74438bddc22ff1464760276078402a41a" gracePeriod=30 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.426866 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="proxy-httpd" containerID="cri-o://3b33a495951843dca3306d1f5afb9c903071ebc4d8c7ab9493eea6fb248bc2b5" gracePeriod=30 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974314 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerID="3b33a495951843dca3306d1f5afb9c903071ebc4d8c7ab9493eea6fb248bc2b5" exitCode=0 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974697 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerID="b8c91769273de4970f74a2edfeccc49f78d58484cf0ceb4beb6cff90550e93be" exitCode=2 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974712 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerID="ffbcb2b2210cf93d25de2e835c74755e1cff96ca38e2c6ac4b7c6dceea54332d" exitCode=0 Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974480 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerDied","Data":"3b33a495951843dca3306d1f5afb9c903071ebc4d8c7ab9493eea6fb248bc2b5"} Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974756 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerDied","Data":"b8c91769273de4970f74a2edfeccc49f78d58484cf0ceb4beb6cff90550e93be"} Oct 14 08:59:04 crc kubenswrapper[5018]: I1014 08:59:04.974774 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerDied","Data":"ffbcb2b2210cf93d25de2e835c74755e1cff96ca38e2c6ac4b7c6dceea54332d"} Oct 14 08:59:05 crc kubenswrapper[5018]: E1014 08:59:05.014930 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.436773 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.525005 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd8tl\" (UniqueName: \"kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl\") pod \"5174f11a-5230-4d98-b393-82a90852f9dd\" (UID: \"5174f11a-5230-4d98-b393-82a90852f9dd\") " Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.530480 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl" (OuterVolumeSpecName: "kube-api-access-gd8tl") pod "5174f11a-5230-4d98-b393-82a90852f9dd" (UID: "5174f11a-5230-4d98-b393-82a90852f9dd"). InnerVolumeSpecName "kube-api-access-gd8tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.628680 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd8tl\" (UniqueName: \"kubernetes.io/projected/5174f11a-5230-4d98-b393-82a90852f9dd-kube-api-access-gd8tl\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.986219 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 08:59:05 crc kubenswrapper[5018]: E1014 08:59:05.986472 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.990232 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.990238 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5174f11a-5230-4d98-b393-82a90852f9dd","Type":"ContainerDied","Data":"4bd8e9599e84c24b16e0562c61e849c97e51f526de4c5b315646ea1b44fb12c6"} Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.990286 5018 scope.go:117] "RemoveContainer" containerID="25666f73a516304a499c435a4de6d90f10c044eb527d649f6a7c5c95162f6ffc" Oct 14 08:59:05 crc kubenswrapper[5018]: I1014 08:59:05.999102 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-s48hc" event={"ID":"208ccbe0-ffd5-478f-aca6-8af7d38f9b33","Type":"ContainerStarted","Data":"b301967d4689fd683a3b9fe925bc51719be3a31512b8b531cf8d4a04678a438c"} Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.019104 5018 generic.go:334] "Generic (PLEG): container finished" podID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerID="71b14d6647d11fec0ae387faf95137d74438bddc22ff1464760276078402a41a" exitCode=0 Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.019355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerDied","Data":"71b14d6647d11fec0ae387faf95137d74438bddc22ff1464760276078402a41a"} Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.039488 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-s48hc" podStartSLOduration=2.33817501 podStartE2EDuration="8.039472631s" podCreationTimestamp="2025-10-14 08:58:58 +0000 UTC" firstStartedPulling="2025-10-14 08:58:59.372771859 +0000 UTC m=+7755.956818486" lastFinishedPulling="2025-10-14 08:59:05.07406948 +0000 UTC m=+7761.658116107" observedRunningTime="2025-10-14 08:59:06.029282131 +0000 UTC m=+7762.613328758" watchObservedRunningTime="2025-10-14 08:59:06.039472631 +0000 UTC m=+7762.623519258" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.103124 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.130129 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.137611 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:06 crc kubenswrapper[5018]: E1014 08:59:06.138069 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5174f11a-5230-4d98-b393-82a90852f9dd" containerName="kube-state-metrics" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.138087 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5174f11a-5230-4d98-b393-82a90852f9dd" containerName="kube-state-metrics" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.138302 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5174f11a-5230-4d98-b393-82a90852f9dd" containerName="kube-state-metrics" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.139214 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.141061 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.141060 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.152946 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.241764 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.242154 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrdp2\" (UniqueName: \"kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.242228 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.242294 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.346136 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.346206 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrdp2\" (UniqueName: \"kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.346283 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.346368 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.352909 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.353054 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.353069 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.364109 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrdp2\" (UniqueName: \"kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2\") pod \"kube-state-metrics-0\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.423818 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.468582 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.551591 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.551803 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.551823 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.552098 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.552195 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.552338 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.552453 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vfn8\" (UniqueName: \"kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8\") pod \"fbe08582-ed39-4651-9e4a-393dc07368ec\" (UID: \"fbe08582-ed39-4651-9e4a-393dc07368ec\") " Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.552842 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.553045 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.557012 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8" (OuterVolumeSpecName: "kube-api-access-8vfn8") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "kube-api-access-8vfn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.557553 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.560796 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts" (OuterVolumeSpecName: "scripts") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.586841 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.628809 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5174f11a-5230-4d98-b393-82a90852f9dd" path="/var/lib/kubelet/pods/5174f11a-5230-4d98-b393-82a90852f9dd/volumes" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.654825 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbe08582-ed39-4651-9e4a-393dc07368ec-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.655042 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vfn8\" (UniqueName: \"kubernetes.io/projected/fbe08582-ed39-4651-9e4a-393dc07368ec-kube-api-access-8vfn8\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.655057 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.655065 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.661300 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data" (OuterVolumeSpecName: "config-data") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.661385 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbe08582-ed39-4651-9e4a-393dc07368ec" (UID: "fbe08582-ed39-4651-9e4a-393dc07368ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.756768 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.756803 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe08582-ed39-4651-9e4a-393dc07368ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:06 crc kubenswrapper[5018]: I1014 08:59:06.909982 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:59:06 crc kubenswrapper[5018]: W1014 08:59:06.915319 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d48b611_d8d9_48a6_9179_d58ffa16ac18.slice/crio-5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2 WatchSource:0}: Error finding container 5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2: Status 404 returned error can't find the container with id 5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2 Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.048895 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbe08582-ed39-4651-9e4a-393dc07368ec","Type":"ContainerDied","Data":"94f10b267613730106b78ac8c9f439ff91aa3c3f6b7026086b5592c2067d2b95"} Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.048927 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.049378 5018 scope.go:117] "RemoveContainer" containerID="3b33a495951843dca3306d1f5afb9c903071ebc4d8c7ab9493eea6fb248bc2b5" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.059732 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d48b611-d8d9-48a6-9179-d58ffa16ac18","Type":"ContainerStarted","Data":"5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2"} Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.097100 5018 scope.go:117] "RemoveContainer" containerID="b8c91769273de4970f74a2edfeccc49f78d58484cf0ceb4beb6cff90550e93be" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.109993 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.140703 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.143020 5018 scope.go:117] "RemoveContainer" containerID="71b14d6647d11fec0ae387faf95137d74438bddc22ff1464760276078402a41a" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.152377 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:07 crc kubenswrapper[5018]: E1014 08:59:07.152873 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-central-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.152893 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-central-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: E1014 08:59:07.152910 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-notification-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.152917 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-notification-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: E1014 08:59:07.152947 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="sg-core" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.152954 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="sg-core" Oct 14 08:59:07 crc kubenswrapper[5018]: E1014 08:59:07.152963 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="proxy-httpd" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.152969 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="proxy-httpd" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.153153 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-notification-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.153171 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="sg-core" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.153178 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="proxy-httpd" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.153192 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" containerName="ceilometer-central-agent" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.155556 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.158485 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.158900 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.159068 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.173901 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.179535 5018 scope.go:117] "RemoveContainer" containerID="ffbcb2b2210cf93d25de2e835c74755e1cff96ca38e2c6ac4b7c6dceea54332d" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.266983 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267122 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267485 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267516 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267531 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267677 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkbgx\" (UniqueName: \"kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267792 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.267890 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370257 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370311 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370336 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370371 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkbgx\" (UniqueName: \"kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370412 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370458 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370510 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.370548 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.371049 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.371044 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.374971 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.376006 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.376271 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.376604 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.378297 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.390522 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkbgx\" (UniqueName: \"kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx\") pod \"ceilometer-0\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.476145 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:07 crc kubenswrapper[5018]: I1014 08:59:07.984553 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:07 crc kubenswrapper[5018]: W1014 08:59:07.994122 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ba2a6ae_95e4_4473_b2f7_0ad7ccc7b7d5.slice/crio-b87bf74297f275c45836d06cee370e927840e4bf415eb37d21c9dbf95da2c7e0 WatchSource:0}: Error finding container b87bf74297f275c45836d06cee370e927840e4bf415eb37d21c9dbf95da2c7e0: Status 404 returned error can't find the container with id b87bf74297f275c45836d06cee370e927840e4bf415eb37d21c9dbf95da2c7e0 Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.008738 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.026695 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.076602 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerStarted","Data":"b87bf74297f275c45836d06cee370e927840e4bf415eb37d21c9dbf95da2c7e0"} Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.080824 5018 generic.go:334] "Generic (PLEG): container finished" podID="208ccbe0-ffd5-478f-aca6-8af7d38f9b33" containerID="b301967d4689fd683a3b9fe925bc51719be3a31512b8b531cf8d4a04678a438c" exitCode=0 Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.080903 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-s48hc" event={"ID":"208ccbe0-ffd5-478f-aca6-8af7d38f9b33","Type":"ContainerDied","Data":"b301967d4689fd683a3b9fe925bc51719be3a31512b8b531cf8d4a04678a438c"} Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.084928 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d48b611-d8d9-48a6-9179-d58ffa16ac18","Type":"ContainerStarted","Data":"0835e3a86cc1bed0b46ea89d60d79c970f2f5d7aa91a959459086c75a2a30e00"} Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.085219 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.098124 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.125077 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.745500552 podStartE2EDuration="2.125055271s" podCreationTimestamp="2025-10-14 08:59:06 +0000 UTC" firstStartedPulling="2025-10-14 08:59:06.919305388 +0000 UTC m=+7763.503352015" lastFinishedPulling="2025-10-14 08:59:07.298860087 +0000 UTC m=+7763.882906734" observedRunningTime="2025-10-14 08:59:08.114109249 +0000 UTC m=+7764.698155886" watchObservedRunningTime="2025-10-14 08:59:08.125055271 +0000 UTC m=+7764.709101898" Oct 14 08:59:08 crc kubenswrapper[5018]: I1014 08:59:08.617559 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbe08582-ed39-4651-9e4a-393dc07368ec" path="/var/lib/kubelet/pods/fbe08582-ed39-4651-9e4a-393dc07368ec/volumes" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.108055 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerStarted","Data":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.108339 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerStarted","Data":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.591790 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-s48hc" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.722605 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data\") pod \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.722852 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts\") pod \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.722893 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh77v\" (UniqueName: \"kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v\") pod \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.722965 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle\") pod \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\" (UID: \"208ccbe0-ffd5-478f-aca6-8af7d38f9b33\") " Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.728688 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v" (OuterVolumeSpecName: "kube-api-access-rh77v") pod "208ccbe0-ffd5-478f-aca6-8af7d38f9b33" (UID: "208ccbe0-ffd5-478f-aca6-8af7d38f9b33"). InnerVolumeSpecName "kube-api-access-rh77v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.755593 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "208ccbe0-ffd5-478f-aca6-8af7d38f9b33" (UID: "208ccbe0-ffd5-478f-aca6-8af7d38f9b33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.756022 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts" (OuterVolumeSpecName: "scripts") pod "208ccbe0-ffd5-478f-aca6-8af7d38f9b33" (UID: "208ccbe0-ffd5-478f-aca6-8af7d38f9b33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.763658 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data" (OuterVolumeSpecName: "config-data") pod "208ccbe0-ffd5-478f-aca6-8af7d38f9b33" (UID: "208ccbe0-ffd5-478f-aca6-8af7d38f9b33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.825576 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.825648 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.825671 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh77v\" (UniqueName: \"kubernetes.io/projected/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-kube-api-access-rh77v\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:09 crc kubenswrapper[5018]: I1014 08:59:09.825690 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ccbe0-ffd5-478f-aca6-8af7d38f9b33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:10 crc kubenswrapper[5018]: I1014 08:59:10.121694 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerStarted","Data":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} Oct 14 08:59:10 crc kubenswrapper[5018]: I1014 08:59:10.123441 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-s48hc" event={"ID":"208ccbe0-ffd5-478f-aca6-8af7d38f9b33","Type":"ContainerDied","Data":"fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807"} Oct 14 08:59:10 crc kubenswrapper[5018]: I1014 08:59:10.123468 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe36c264265a93bb2d512c4c8d2e74e9e1695c29cea4611ed24fcfb3d3616807" Oct 14 08:59:10 crc kubenswrapper[5018]: I1014 08:59:10.123525 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-s48hc" Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.038330 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ttc4f"] Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.047472 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ttc4f"] Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.148088 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerStarted","Data":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.148298 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.180167 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.15127572 podStartE2EDuration="5.180151922s" podCreationTimestamp="2025-10-14 08:59:07 +0000 UTC" firstStartedPulling="2025-10-14 08:59:07.997390912 +0000 UTC m=+7764.581437549" lastFinishedPulling="2025-10-14 08:59:11.026267124 +0000 UTC m=+7767.610313751" observedRunningTime="2025-10-14 08:59:12.17445904 +0000 UTC m=+7768.758505707" watchObservedRunningTime="2025-10-14 08:59:12.180151922 +0000 UTC m=+7768.764198549" Oct 14 08:59:12 crc kubenswrapper[5018]: I1014 08:59:12.633722 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d53c1c9-9ebf-4356-b632-7152d87b8055" path="/var/lib/kubelet/pods/9d53c1c9-9ebf-4356-b632-7152d87b8055/volumes" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.709444 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:13 crc kubenswrapper[5018]: E1014 08:59:13.711166 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208ccbe0-ffd5-478f-aca6-8af7d38f9b33" containerName="aodh-db-sync" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.711373 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="208ccbe0-ffd5-478f-aca6-8af7d38f9b33" containerName="aodh-db-sync" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.711842 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="208ccbe0-ffd5-478f-aca6-8af7d38f9b33" containerName="aodh-db-sync" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.743959 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.748113 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-ps45p" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.748388 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.749223 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.757786 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.871280 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.871583 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.871723 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjmfk\" (UniqueName: \"kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.871838 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.973300 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.973691 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.973855 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.973965 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjmfk\" (UniqueName: \"kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.980831 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.981038 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.988824 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:13 crc kubenswrapper[5018]: I1014 08:59:13.999539 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjmfk\" (UniqueName: \"kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk\") pod \"aodh-0\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " pod="openstack/aodh-0" Oct 14 08:59:14 crc kubenswrapper[5018]: I1014 08:59:14.090527 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:14 crc kubenswrapper[5018]: I1014 08:59:14.649979 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:15 crc kubenswrapper[5018]: I1014 08:59:15.147128 5018 scope.go:117] "RemoveContainer" containerID="3e0af616afa486a9141c3753b888e9b5377cd72aed3a6a0eccbee9d8d201c317" Oct 14 08:59:15 crc kubenswrapper[5018]: I1014 08:59:15.192259 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerStarted","Data":"4c91a7ead63488664fbddca322097da7487b0cd0fde978bb803fd74e5586759c"} Oct 14 08:59:15 crc kubenswrapper[5018]: I1014 08:59:15.382124 5018 scope.go:117] "RemoveContainer" containerID="df8465136d92b044fec3c7f76559ba6e5b0c147ecfd89e58a800587269160316" Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.097132 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.097451 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-central-agent" containerID="cri-o://59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" gracePeriod=30 Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.097757 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="proxy-httpd" containerID="cri-o://3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" gracePeriod=30 Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.097956 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="sg-core" containerID="cri-o://dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" gracePeriod=30 Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.098021 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-notification-agent" containerID="cri-o://c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" gracePeriod=30 Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.204485 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerStarted","Data":"10ea6304fb5ac010339e3324db27b0c137c67f17ab90bcea9ecfed1e0cc68529"} Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.480931 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 08:59:16 crc kubenswrapper[5018]: I1014 08:59:16.753586 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.154341 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.222447 5018 generic.go:334] "Generic (PLEG): container finished" podID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" exitCode=0 Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.226072 5018 generic.go:334] "Generic (PLEG): container finished" podID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" exitCode=2 Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.227674 5018 generic.go:334] "Generic (PLEG): container finished" podID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" exitCode=0 Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.227759 5018 generic.go:334] "Generic (PLEG): container finished" podID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" exitCode=0 Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.227837 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerDied","Data":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.227937 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerDied","Data":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.228001 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerDied","Data":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.228070 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerDied","Data":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.228129 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5","Type":"ContainerDied","Data":"b87bf74297f275c45836d06cee370e927840e4bf415eb37d21c9dbf95da2c7e0"} Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.228204 5018 scope.go:117] "RemoveContainer" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.229508 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.266760 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.266960 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267002 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267033 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkbgx\" (UniqueName: \"kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267077 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267111 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267175 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts\") pod \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\" (UID: \"9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5\") " Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267435 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.267508 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.268154 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.268176 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.272574 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts" (OuterVolumeSpecName: "scripts") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.272654 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx" (OuterVolumeSpecName: "kube-api-access-kkbgx") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "kube-api-access-kkbgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.310795 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.331831 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.369837 5018 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.369871 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkbgx\" (UniqueName: \"kubernetes.io/projected/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-kube-api-access-kkbgx\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.369886 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.369899 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.370892 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.388956 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data" (OuterVolumeSpecName: "config-data") pod "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" (UID: "9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.457157 5018 scope.go:117] "RemoveContainer" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.471538 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.471571 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.474461 5018 scope.go:117] "RemoveContainer" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.499615 5018 scope.go:117] "RemoveContainer" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.525295 5018 scope.go:117] "RemoveContainer" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.525884 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": container with ID starting with 3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00 not found: ID does not exist" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.525931 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} err="failed to get container status \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": rpc error: code = NotFound desc = could not find container \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": container with ID starting with 3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.525963 5018 scope.go:117] "RemoveContainer" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.526306 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": container with ID starting with dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03 not found: ID does not exist" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.526341 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} err="failed to get container status \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": rpc error: code = NotFound desc = could not find container \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": container with ID starting with dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.526364 5018 scope.go:117] "RemoveContainer" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.526900 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": container with ID starting with c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b not found: ID does not exist" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.526940 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} err="failed to get container status \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": rpc error: code = NotFound desc = could not find container \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": container with ID starting with c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.526963 5018 scope.go:117] "RemoveContainer" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.527231 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": container with ID starting with 59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf not found: ID does not exist" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.527270 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} err="failed to get container status \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": rpc error: code = NotFound desc = could not find container \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": container with ID starting with 59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.527295 5018 scope.go:117] "RemoveContainer" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.527538 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} err="failed to get container status \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": rpc error: code = NotFound desc = could not find container \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": container with ID starting with 3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.527570 5018 scope.go:117] "RemoveContainer" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.528057 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} err="failed to get container status \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": rpc error: code = NotFound desc = could not find container \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": container with ID starting with dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.528090 5018 scope.go:117] "RemoveContainer" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.528597 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} err="failed to get container status \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": rpc error: code = NotFound desc = could not find container \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": container with ID starting with c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.528645 5018 scope.go:117] "RemoveContainer" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.528994 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} err="failed to get container status \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": rpc error: code = NotFound desc = could not find container \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": container with ID starting with 59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.529026 5018 scope.go:117] "RemoveContainer" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.529381 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} err="failed to get container status \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": rpc error: code = NotFound desc = could not find container \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": container with ID starting with 3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.529409 5018 scope.go:117] "RemoveContainer" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.529735 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} err="failed to get container status \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": rpc error: code = NotFound desc = could not find container \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": container with ID starting with dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.529771 5018 scope.go:117] "RemoveContainer" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.531825 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} err="failed to get container status \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": rpc error: code = NotFound desc = could not find container \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": container with ID starting with c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.531958 5018 scope.go:117] "RemoveContainer" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.532441 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} err="failed to get container status \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": rpc error: code = NotFound desc = could not find container \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": container with ID starting with 59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.532506 5018 scope.go:117] "RemoveContainer" containerID="3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.532959 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00"} err="failed to get container status \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": rpc error: code = NotFound desc = could not find container \"3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00\": container with ID starting with 3a568038a5bbe59fc5e0582e3f25e38eef8d84ede8af6d0f01d1713de5ae5a00 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.532997 5018 scope.go:117] "RemoveContainer" containerID="dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.533327 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03"} err="failed to get container status \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": rpc error: code = NotFound desc = could not find container \"dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03\": container with ID starting with dd42910baae6cdd12b00d87ec545f85c678faa0c818f4a589fcabd7838c43a03 not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.533361 5018 scope.go:117] "RemoveContainer" containerID="c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.533778 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b"} err="failed to get container status \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": rpc error: code = NotFound desc = could not find container \"c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b\": container with ID starting with c6e0c542e9156eddc24df78a9f97fd9362bceacb23fde7e1f1f30fed3fa36f8b not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.533827 5018 scope.go:117] "RemoveContainer" containerID="59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.534198 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf"} err="failed to get container status \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": rpc error: code = NotFound desc = could not find container \"59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf\": container with ID starting with 59a05fc363527e60ccd753591ba3289846a9a5a993177c53fba05b445aad1dcf not found: ID does not exist" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.576713 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.601545 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.614639 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.615113 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="proxy-httpd" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615129 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="proxy-httpd" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.615147 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="sg-core" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615158 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="sg-core" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.615182 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-central-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615190 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-central-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: E1014 08:59:17.615232 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-notification-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615241 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-notification-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615820 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="sg-core" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615841 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="proxy-httpd" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615861 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-central-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.615886 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" containerName="ceilometer-notification-agent" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.618803 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.622405 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.622525 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.622674 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.628055 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675113 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675175 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675280 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675429 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675577 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gph9d\" (UniqueName: \"kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675739 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.675783 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.676022 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778350 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778405 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gph9d\" (UniqueName: \"kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778462 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778493 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778558 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778728 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778769 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.778794 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.780134 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.780338 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.784186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.784243 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.784501 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.795137 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.802043 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gph9d\" (UniqueName: \"kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.805282 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " pod="openstack/ceilometer-0" Oct 14 08:59:17 crc kubenswrapper[5018]: I1014 08:59:17.954073 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:18 crc kubenswrapper[5018]: I1014 08:59:18.175878 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:18 crc kubenswrapper[5018]: I1014 08:59:18.264141 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerStarted","Data":"e75da823fed7de7b31f4d43e999ab665655abab1201befd87f9e5a24f0190080"} Oct 14 08:59:18 crc kubenswrapper[5018]: I1014 08:59:18.436420 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:18 crc kubenswrapper[5018]: I1014 08:59:18.606563 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 08:59:18 crc kubenswrapper[5018]: E1014 08:59:18.607071 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:18 crc kubenswrapper[5018]: W1014 08:59:18.613410 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f524990_21af_4831_9b17_e55cd90c87b5.slice/crio-7fd5f10cedb7f33949036a84273208e7551c6cd301758ce46211c1cc19902084 WatchSource:0}: Error finding container 7fd5f10cedb7f33949036a84273208e7551c6cd301758ce46211c1cc19902084: Status 404 returned error can't find the container with id 7fd5f10cedb7f33949036a84273208e7551c6cd301758ce46211c1cc19902084 Oct 14 08:59:18 crc kubenswrapper[5018]: I1014 08:59:18.620877 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5" path="/var/lib/kubelet/pods/9ba2a6ae-95e4-4473-b2f7-0ad7ccc7b7d5/volumes" Oct 14 08:59:19 crc kubenswrapper[5018]: I1014 08:59:19.298245 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerStarted","Data":"8823c0aa7e7bb68f4ee23a6cd899e92e191efc1e08d1b2aece76aeb94e6bb38b"} Oct 14 08:59:19 crc kubenswrapper[5018]: I1014 08:59:19.298757 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerStarted","Data":"7fd5f10cedb7f33949036a84273208e7551c6cd301758ce46211c1cc19902084"} Oct 14 08:59:19 crc kubenswrapper[5018]: I1014 08:59:19.303871 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerStarted","Data":"3ff4e53d026046161fc0ba30442e5f12b23e128b7dfd5eb8afd92d80365c87b0"} Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.314456 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerStarted","Data":"2e18282fc72e88490b59a6305aca50849fe44120a6854b55dfe67dd97d12eb94"} Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.314837 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerStarted","Data":"b7ed81f42946d37c8a553404dda41fd30f4cab67ddca9684412c6480254a5ba4"} Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.317419 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerStarted","Data":"a37980b420c8b815e40b9a904710947c3fc4d3660e112850bdb5cb237e5eafaf"} Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.317589 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-api" containerID="cri-o://10ea6304fb5ac010339e3324db27b0c137c67f17ab90bcea9ecfed1e0cc68529" gracePeriod=30 Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.318233 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-listener" containerID="cri-o://a37980b420c8b815e40b9a904710947c3fc4d3660e112850bdb5cb237e5eafaf" gracePeriod=30 Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.318299 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-notifier" containerID="cri-o://3ff4e53d026046161fc0ba30442e5f12b23e128b7dfd5eb8afd92d80365c87b0" gracePeriod=30 Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.318372 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-evaluator" containerID="cri-o://e75da823fed7de7b31f4d43e999ab665655abab1201befd87f9e5a24f0190080" gracePeriod=30 Oct 14 08:59:20 crc kubenswrapper[5018]: I1014 08:59:20.348823 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.099732837 podStartE2EDuration="7.348797325s" podCreationTimestamp="2025-10-14 08:59:13 +0000 UTC" firstStartedPulling="2025-10-14 08:59:14.665405912 +0000 UTC m=+7771.249452539" lastFinishedPulling="2025-10-14 08:59:19.9144704 +0000 UTC m=+7776.498517027" observedRunningTime="2025-10-14 08:59:20.344451051 +0000 UTC m=+7776.928497678" watchObservedRunningTime="2025-10-14 08:59:20.348797325 +0000 UTC m=+7776.932843952" Oct 14 08:59:21 crc kubenswrapper[5018]: I1014 08:59:21.333194 5018 generic.go:334] "Generic (PLEG): container finished" podID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerID="e75da823fed7de7b31f4d43e999ab665655abab1201befd87f9e5a24f0190080" exitCode=0 Oct 14 08:59:21 crc kubenswrapper[5018]: I1014 08:59:21.333567 5018 generic.go:334] "Generic (PLEG): container finished" podID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerID="10ea6304fb5ac010339e3324db27b0c137c67f17ab90bcea9ecfed1e0cc68529" exitCode=0 Oct 14 08:59:21 crc kubenswrapper[5018]: I1014 08:59:21.333282 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerDied","Data":"e75da823fed7de7b31f4d43e999ab665655abab1201befd87f9e5a24f0190080"} Oct 14 08:59:21 crc kubenswrapper[5018]: I1014 08:59:21.333636 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerDied","Data":"10ea6304fb5ac010339e3324db27b0c137c67f17ab90bcea9ecfed1e0cc68529"} Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.045701 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-72c6-account-create-5hkcx"] Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.064171 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-72c6-account-create-5hkcx"] Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345094 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerStarted","Data":"1b36bb8cd97eae0d74c747cd1c68f0d49f3f2b30bfb1e7ce00d71d8ccc4fe875"} Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345338 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345346 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="proxy-httpd" containerID="cri-o://1b36bb8cd97eae0d74c747cd1c68f0d49f3f2b30bfb1e7ce00d71d8ccc4fe875" gracePeriod=30 Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345365 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-notification-agent" containerID="cri-o://2e18282fc72e88490b59a6305aca50849fe44120a6854b55dfe67dd97d12eb94" gracePeriod=30 Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345347 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="sg-core" containerID="cri-o://b7ed81f42946d37c8a553404dda41fd30f4cab67ddca9684412c6480254a5ba4" gracePeriod=30 Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.345466 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-central-agent" containerID="cri-o://8823c0aa7e7bb68f4ee23a6cd899e92e191efc1e08d1b2aece76aeb94e6bb38b" gracePeriod=30 Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.372007 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.671288337 podStartE2EDuration="5.371991331s" podCreationTimestamp="2025-10-14 08:59:17 +0000 UTC" firstStartedPulling="2025-10-14 08:59:18.60854179 +0000 UTC m=+7775.192588417" lastFinishedPulling="2025-10-14 08:59:21.309244764 +0000 UTC m=+7777.893291411" observedRunningTime="2025-10-14 08:59:22.367787782 +0000 UTC m=+7778.951834409" watchObservedRunningTime="2025-10-14 08:59:22.371991331 +0000 UTC m=+7778.956037958" Oct 14 08:59:22 crc kubenswrapper[5018]: I1014 08:59:22.615808 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45fc70f1-8084-4a78-bbd1-0ee01009bdd8" path="/var/lib/kubelet/pods/45fc70f1-8084-4a78-bbd1-0ee01009bdd8/volumes" Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.372869 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerDied","Data":"1b36bb8cd97eae0d74c747cd1c68f0d49f3f2b30bfb1e7ce00d71d8ccc4fe875"} Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.372817 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f524990-21af-4831-9b17-e55cd90c87b5" containerID="1b36bb8cd97eae0d74c747cd1c68f0d49f3f2b30bfb1e7ce00d71d8ccc4fe875" exitCode=0 Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.372979 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f524990-21af-4831-9b17-e55cd90c87b5" containerID="b7ed81f42946d37c8a553404dda41fd30f4cab67ddca9684412c6480254a5ba4" exitCode=2 Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.372995 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f524990-21af-4831-9b17-e55cd90c87b5" containerID="2e18282fc72e88490b59a6305aca50849fe44120a6854b55dfe67dd97d12eb94" exitCode=0 Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.373025 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerDied","Data":"b7ed81f42946d37c8a553404dda41fd30f4cab67ddca9684412c6480254a5ba4"} Oct 14 08:59:23 crc kubenswrapper[5018]: I1014 08:59:23.373116 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerDied","Data":"2e18282fc72e88490b59a6305aca50849fe44120a6854b55dfe67dd97d12eb94"} Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.398830 5018 generic.go:334] "Generic (PLEG): container finished" podID="0f524990-21af-4831-9b17-e55cd90c87b5" containerID="8823c0aa7e7bb68f4ee23a6cd899e92e191efc1e08d1b2aece76aeb94e6bb38b" exitCode=0 Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.399094 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerDied","Data":"8823c0aa7e7bb68f4ee23a6cd899e92e191efc1e08d1b2aece76aeb94e6bb38b"} Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.752426 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.937757 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.937915 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.937954 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.937998 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.938093 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gph9d\" (UniqueName: \"kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.938206 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.938359 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.938417 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml\") pod \"0f524990-21af-4831-9b17-e55cd90c87b5\" (UID: \"0f524990-21af-4831-9b17-e55cd90c87b5\") " Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.938417 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.939136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.939399 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.939442 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f524990-21af-4831-9b17-e55cd90c87b5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.945742 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d" (OuterVolumeSpecName: "kube-api-access-gph9d") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "kube-api-access-gph9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.947302 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts" (OuterVolumeSpecName: "scripts") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:24 crc kubenswrapper[5018]: I1014 08:59:24.974608 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.041423 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.042520 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gph9d\" (UniqueName: \"kubernetes.io/projected/0f524990-21af-4831-9b17-e55cd90c87b5-kube-api-access-gph9d\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.042559 5018 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.042598 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.042672 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.091895 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.130548 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data" (OuterVolumeSpecName: "config-data") pod "0f524990-21af-4831-9b17-e55cd90c87b5" (UID: "0f524990-21af-4831-9b17-e55cd90c87b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.145703 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.145755 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f524990-21af-4831-9b17-e55cd90c87b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.423271 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f524990-21af-4831-9b17-e55cd90c87b5","Type":"ContainerDied","Data":"7fd5f10cedb7f33949036a84273208e7551c6cd301758ce46211c1cc19902084"} Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.423354 5018 scope.go:117] "RemoveContainer" containerID="1b36bb8cd97eae0d74c747cd1c68f0d49f3f2b30bfb1e7ce00d71d8ccc4fe875" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.423390 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.459775 5018 scope.go:117] "RemoveContainer" containerID="b7ed81f42946d37c8a553404dda41fd30f4cab67ddca9684412c6480254a5ba4" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.493408 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.506533 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.519244 5018 scope.go:117] "RemoveContainer" containerID="2e18282fc72e88490b59a6305aca50849fe44120a6854b55dfe67dd97d12eb94" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.531491 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:25 crc kubenswrapper[5018]: E1014 08:59:25.531945 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="proxy-httpd" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.531968 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="proxy-httpd" Oct 14 08:59:25 crc kubenswrapper[5018]: E1014 08:59:25.531984 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="sg-core" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.531992 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="sg-core" Oct 14 08:59:25 crc kubenswrapper[5018]: E1014 08:59:25.532040 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-notification-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532049 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-notification-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: E1014 08:59:25.532063 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-central-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532071 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-central-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532295 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-central-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532311 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="proxy-httpd" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532335 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="sg-core" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.532348 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" containerName="ceilometer-notification-agent" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.549159 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.549288 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.552800 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.553127 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.553310 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556380 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556448 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556534 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmhn6\" (UniqueName: \"kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556581 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556604 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556642 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556661 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.556701 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.557011 5018 scope.go:117] "RemoveContainer" containerID="8823c0aa7e7bb68f4ee23a6cd899e92e191efc1e08d1b2aece76aeb94e6bb38b" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.658847 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmhn6\" (UniqueName: \"kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.659198 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.659313 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.659426 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.659730 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.659879 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.660044 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.660205 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.661104 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.661760 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.665424 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.665882 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.666433 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.666713 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.675979 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.688986 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmhn6\" (UniqueName: \"kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6\") pod \"ceilometer-0\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " pod="openstack/ceilometer-0" Oct 14 08:59:25 crc kubenswrapper[5018]: I1014 08:59:25.874906 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:59:26 crc kubenswrapper[5018]: I1014 08:59:26.416109 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:59:26 crc kubenswrapper[5018]: W1014 08:59:26.439787 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf09a285_4bbe_4a1a_9fa2_9aa1ec047328.slice/crio-62ae37e1c397e59bdf1a805b4423f2e0bf8bf212741bb9d0d6f8fc36b878d0a7 WatchSource:0}: Error finding container 62ae37e1c397e59bdf1a805b4423f2e0bf8bf212741bb9d0d6f8fc36b878d0a7: Status 404 returned error can't find the container with id 62ae37e1c397e59bdf1a805b4423f2e0bf8bf212741bb9d0d6f8fc36b878d0a7 Oct 14 08:59:26 crc kubenswrapper[5018]: I1014 08:59:26.621570 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f524990-21af-4831-9b17-e55cd90c87b5" path="/var/lib/kubelet/pods/0f524990-21af-4831-9b17-e55cd90c87b5/volumes" Oct 14 08:59:27 crc kubenswrapper[5018]: I1014 08:59:27.460776 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerStarted","Data":"298b04d7b493d03a2ba6b0a665cc5d3190695959aea3d6fec443b33a4c7ab1c7"} Oct 14 08:59:27 crc kubenswrapper[5018]: I1014 08:59:27.461131 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerStarted","Data":"68ece5d157ffc84bb5586ef4fa04f0d496bf0dbee2e3cc8cbcd837a81d450889"} Oct 14 08:59:27 crc kubenswrapper[5018]: I1014 08:59:27.461143 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerStarted","Data":"62ae37e1c397e59bdf1a805b4423f2e0bf8bf212741bb9d0d6f8fc36b878d0a7"} Oct 14 08:59:28 crc kubenswrapper[5018]: I1014 08:59:28.496261 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerStarted","Data":"087e0937d73febf55e6d268d5a2a924ae960ed5b8f8408710ab27cc2646a8605"} Oct 14 08:59:29 crc kubenswrapper[5018]: I1014 08:59:29.517091 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerStarted","Data":"82fa2b51a89e84f109bc39c3a296ac0d8bc72f1de856de23122ba2cda3338fdd"} Oct 14 08:59:29 crc kubenswrapper[5018]: I1014 08:59:29.517939 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:59:29 crc kubenswrapper[5018]: I1014 08:59:29.555465 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.006686818 podStartE2EDuration="4.555441763s" podCreationTimestamp="2025-10-14 08:59:25 +0000 UTC" firstStartedPulling="2025-10-14 08:59:26.443578212 +0000 UTC m=+7783.027624839" lastFinishedPulling="2025-10-14 08:59:28.992333157 +0000 UTC m=+7785.576379784" observedRunningTime="2025-10-14 08:59:29.553832437 +0000 UTC m=+7786.137879104" watchObservedRunningTime="2025-10-14 08:59:29.555441763 +0000 UTC m=+7786.139488420" Oct 14 08:59:29 crc kubenswrapper[5018]: I1014 08:59:29.605849 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 08:59:29 crc kubenswrapper[5018]: E1014 08:59:29.606307 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:40 crc kubenswrapper[5018]: I1014 08:59:40.606116 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 08:59:40 crc kubenswrapper[5018]: E1014 08:59:40.607595 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:46 crc kubenswrapper[5018]: I1014 08:59:46.067601 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-l4hcx"] Oct 14 08:59:46 crc kubenswrapper[5018]: I1014 08:59:46.082488 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-l4hcx"] Oct 14 08:59:46 crc kubenswrapper[5018]: I1014 08:59:46.623043 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b57a52e6-9c24-4925-8503-c5e9cd43de14" path="/var/lib/kubelet/pods/b57a52e6-9c24-4925-8503-c5e9cd43de14/volumes" Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.856510 5018 generic.go:334] "Generic (PLEG): container finished" podID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerID="a37980b420c8b815e40b9a904710947c3fc4d3660e112850bdb5cb237e5eafaf" exitCode=137 Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.857208 5018 generic.go:334] "Generic (PLEG): container finished" podID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerID="3ff4e53d026046161fc0ba30442e5f12b23e128b7dfd5eb8afd92d80365c87b0" exitCode=137 Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.856610 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerDied","Data":"a37980b420c8b815e40b9a904710947c3fc4d3660e112850bdb5cb237e5eafaf"} Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.857241 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerDied","Data":"3ff4e53d026046161fc0ba30442e5f12b23e128b7dfd5eb8afd92d80365c87b0"} Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.857252 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8dbf4810-af23-4f1a-8008-2c34bc7706e5","Type":"ContainerDied","Data":"4c91a7ead63488664fbddca322097da7487b0cd0fde978bb803fd74e5586759c"} Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.857261 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c91a7ead63488664fbddca322097da7487b0cd0fde978bb803fd74e5586759c" Oct 14 08:59:50 crc kubenswrapper[5018]: I1014 08:59:50.918208 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.012295 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjmfk\" (UniqueName: \"kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk\") pod \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.012480 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts\") pod \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.012532 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data\") pod \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.012651 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle\") pod \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\" (UID: \"8dbf4810-af23-4f1a-8008-2c34bc7706e5\") " Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.035135 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts" (OuterVolumeSpecName: "scripts") pod "8dbf4810-af23-4f1a-8008-2c34bc7706e5" (UID: "8dbf4810-af23-4f1a-8008-2c34bc7706e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.036492 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk" (OuterVolumeSpecName: "kube-api-access-jjmfk") pod "8dbf4810-af23-4f1a-8008-2c34bc7706e5" (UID: "8dbf4810-af23-4f1a-8008-2c34bc7706e5"). InnerVolumeSpecName "kube-api-access-jjmfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.115374 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.115411 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjmfk\" (UniqueName: \"kubernetes.io/projected/8dbf4810-af23-4f1a-8008-2c34bc7706e5-kube-api-access-jjmfk\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.152291 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dbf4810-af23-4f1a-8008-2c34bc7706e5" (UID: "8dbf4810-af23-4f1a-8008-2c34bc7706e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.167294 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data" (OuterVolumeSpecName: "config-data") pod "8dbf4810-af23-4f1a-8008-2c34bc7706e5" (UID: "8dbf4810-af23-4f1a-8008-2c34bc7706e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.217337 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.217367 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dbf4810-af23-4f1a-8008-2c34bc7706e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.870042 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.957512 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.973490 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988030 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:51 crc kubenswrapper[5018]: E1014 08:59:51.988487 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-notifier" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988500 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-notifier" Oct 14 08:59:51 crc kubenswrapper[5018]: E1014 08:59:51.988518 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-listener" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988523 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-listener" Oct 14 08:59:51 crc kubenswrapper[5018]: E1014 08:59:51.988547 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-evaluator" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988554 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-evaluator" Oct 14 08:59:51 crc kubenswrapper[5018]: E1014 08:59:51.988568 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-api" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988573 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-api" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988799 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-listener" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988813 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-api" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988829 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-notifier" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.988840 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" containerName="aodh-evaluator" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.990569 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.997586 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-ps45p" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.997788 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.997841 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.997601 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.998003 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 14 08:59:51 crc kubenswrapper[5018]: I1014 08:59:51.998270 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144166 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144298 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144355 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144500 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwlbr\" (UniqueName: \"kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144560 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.144627 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247195 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247322 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247365 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247457 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwlbr\" (UniqueName: \"kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247489 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.247526 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.252817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.253210 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.253817 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.254211 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.256269 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.264186 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwlbr\" (UniqueName: \"kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr\") pod \"aodh-0\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.320900 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.619031 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dbf4810-af23-4f1a-8008-2c34bc7706e5" path="/var/lib/kubelet/pods/8dbf4810-af23-4f1a-8008-2c34bc7706e5/volumes" Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.787147 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 08:59:52 crc kubenswrapper[5018]: W1014 08:59:52.791577 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4849b9e9_47c2_49bc_aa83_c4a0e7e16bec.slice/crio-b12d8b4a9229f30afca9bd014771ef78b595311ae1791a5f1e757fde4319d886 WatchSource:0}: Error finding container b12d8b4a9229f30afca9bd014771ef78b595311ae1791a5f1e757fde4319d886: Status 404 returned error can't find the container with id b12d8b4a9229f30afca9bd014771ef78b595311ae1791a5f1e757fde4319d886 Oct 14 08:59:52 crc kubenswrapper[5018]: I1014 08:59:52.880875 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerStarted","Data":"b12d8b4a9229f30afca9bd014771ef78b595311ae1791a5f1e757fde4319d886"} Oct 14 08:59:53 crc kubenswrapper[5018]: I1014 08:59:53.894000 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerStarted","Data":"a7585a2f1c37963c5dbf63572362ad5333d952de5f762503637a079285e5bd70"} Oct 14 08:59:53 crc kubenswrapper[5018]: I1014 08:59:53.894607 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerStarted","Data":"2d821f28ca619425ae5555a780634062a928277adbe2e77a11f5825e3ed63b9d"} Oct 14 08:59:54 crc kubenswrapper[5018]: I1014 08:59:54.631801 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 08:59:54 crc kubenswrapper[5018]: E1014 08:59:54.633754 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 08:59:54 crc kubenswrapper[5018]: I1014 08:59:54.905735 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerStarted","Data":"84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096"} Oct 14 08:59:54 crc kubenswrapper[5018]: I1014 08:59:54.905983 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerStarted","Data":"4d5980364138e871981e0c2a309c64682bdee2ba69842dd2a7eb988420399d1e"} Oct 14 08:59:54 crc kubenswrapper[5018]: I1014 08:59:54.958229 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.536898195 podStartE2EDuration="3.958206783s" podCreationTimestamp="2025-10-14 08:59:51 +0000 UTC" firstStartedPulling="2025-10-14 08:59:52.794449091 +0000 UTC m=+7809.378495728" lastFinishedPulling="2025-10-14 08:59:54.215757689 +0000 UTC m=+7810.799804316" observedRunningTime="2025-10-14 08:59:54.938020219 +0000 UTC m=+7811.522066846" watchObservedRunningTime="2025-10-14 08:59:54.958206783 +0000 UTC m=+7811.542253410" Oct 14 08:59:55 crc kubenswrapper[5018]: I1014 08:59:55.914437 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.554598 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.557401 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.576805 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.617112 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdjq8\" (UniqueName: \"kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.617233 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.617268 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.719438 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdjq8\" (UniqueName: \"kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.719668 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.719709 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.722349 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.722398 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.747531 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdjq8\" (UniqueName: \"kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8\") pod \"redhat-operators-f8l5t\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:58 crc kubenswrapper[5018]: I1014 08:59:58.941116 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 08:59:59 crc kubenswrapper[5018]: I1014 08:59:59.430785 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 08:59:59 crc kubenswrapper[5018]: W1014 08:59:59.438774 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc8b7d7d_f39b_493f_92ec_e495a46b7b86.slice/crio-1b96d4330d7e61bd75906732a6f8418c8a375f7e35bfce4bffa6ffe8f0faf151 WatchSource:0}: Error finding container 1b96d4330d7e61bd75906732a6f8418c8a375f7e35bfce4bffa6ffe8f0faf151: Status 404 returned error can't find the container with id 1b96d4330d7e61bd75906732a6f8418c8a375f7e35bfce4bffa6ffe8f0faf151 Oct 14 08:59:59 crc kubenswrapper[5018]: I1014 08:59:59.969191 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerID="cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11" exitCode=0 Oct 14 08:59:59 crc kubenswrapper[5018]: I1014 08:59:59.969236 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerDied","Data":"cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11"} Oct 14 08:59:59 crc kubenswrapper[5018]: I1014 08:59:59.969396 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerStarted","Data":"1b96d4330d7e61bd75906732a6f8418c8a375f7e35bfce4bffa6ffe8f0faf151"} Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.145276 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76"] Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.147360 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.148994 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.155393 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.156361 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76"] Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.264186 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.264498 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.264891 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xss6s\" (UniqueName: \"kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.367051 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xss6s\" (UniqueName: \"kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.367223 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.367306 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.368460 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.375369 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.396308 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xss6s\" (UniqueName: \"kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s\") pod \"collect-profiles-29340540-zlm76\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.474092 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:00 crc kubenswrapper[5018]: I1014 09:00:00.981753 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76"] Oct 14 09:00:01 crc kubenswrapper[5018]: I1014 09:00:01.991166 5018 generic.go:334] "Generic (PLEG): container finished" podID="70a68fb7-9603-41e6-a75a-40b35c5c041d" containerID="ae24a91269c1c2e2e51d1eb92dafbd4572d14cdba06bd4208d12998d0cf7af26" exitCode=0 Oct 14 09:00:01 crc kubenswrapper[5018]: I1014 09:00:01.991222 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" event={"ID":"70a68fb7-9603-41e6-a75a-40b35c5c041d","Type":"ContainerDied","Data":"ae24a91269c1c2e2e51d1eb92dafbd4572d14cdba06bd4208d12998d0cf7af26"} Oct 14 09:00:01 crc kubenswrapper[5018]: I1014 09:00:01.991533 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" event={"ID":"70a68fb7-9603-41e6-a75a-40b35c5c041d","Type":"ContainerStarted","Data":"d76b33f7bd158b0d04bf4c70357c9fafeb148db2566d1569fe6a0ddf6fa8903a"} Oct 14 09:00:01 crc kubenswrapper[5018]: I1014 09:00:01.993550 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerStarted","Data":"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335"} Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.379205 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.536039 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume\") pod \"70a68fb7-9603-41e6-a75a-40b35c5c041d\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.536164 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xss6s\" (UniqueName: \"kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s\") pod \"70a68fb7-9603-41e6-a75a-40b35c5c041d\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.536472 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume\") pod \"70a68fb7-9603-41e6-a75a-40b35c5c041d\" (UID: \"70a68fb7-9603-41e6-a75a-40b35c5c041d\") " Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.537098 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume" (OuterVolumeSpecName: "config-volume") pod "70a68fb7-9603-41e6-a75a-40b35c5c041d" (UID: "70a68fb7-9603-41e6-a75a-40b35c5c041d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.537676 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70a68fb7-9603-41e6-a75a-40b35c5c041d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.546413 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "70a68fb7-9603-41e6-a75a-40b35c5c041d" (UID: "70a68fb7-9603-41e6-a75a-40b35c5c041d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.547724 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s" (OuterVolumeSpecName: "kube-api-access-xss6s") pod "70a68fb7-9603-41e6-a75a-40b35c5c041d" (UID: "70a68fb7-9603-41e6-a75a-40b35c5c041d"). InnerVolumeSpecName "kube-api-access-xss6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.639009 5018 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70a68fb7-9603-41e6-a75a-40b35c5c041d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:03 crc kubenswrapper[5018]: I1014 09:00:03.639037 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xss6s\" (UniqueName: \"kubernetes.io/projected/70a68fb7-9603-41e6-a75a-40b35c5c041d-kube-api-access-xss6s\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.012045 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" event={"ID":"70a68fb7-9603-41e6-a75a-40b35c5c041d","Type":"ContainerDied","Data":"d76b33f7bd158b0d04bf4c70357c9fafeb148db2566d1569fe6a0ddf6fa8903a"} Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.012079 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d76b33f7bd158b0d04bf4c70357c9fafeb148db2566d1569fe6a0ddf6fa8903a" Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.012114 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-zlm76" Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.454694 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv"] Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.463267 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-jr2lv"] Oct 14 09:00:04 crc kubenswrapper[5018]: I1014 09:00:04.617540 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5357a32-c284-41a6-b927-b36cc1ee6509" path="/var/lib/kubelet/pods/f5357a32-c284-41a6-b927-b36cc1ee6509/volumes" Oct 14 09:00:05 crc kubenswrapper[5018]: I1014 09:00:05.024070 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerID="d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335" exitCode=0 Oct 14 09:00:05 crc kubenswrapper[5018]: I1014 09:00:05.024106 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerDied","Data":"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335"} Oct 14 09:00:06 crc kubenswrapper[5018]: I1014 09:00:06.605685 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:00:06 crc kubenswrapper[5018]: E1014 09:00:06.606766 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:00:07 crc kubenswrapper[5018]: I1014 09:00:07.051474 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerStarted","Data":"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259"} Oct 14 09:00:07 crc kubenswrapper[5018]: I1014 09:00:07.070153 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f8l5t" podStartSLOduration=2.975940511 podStartE2EDuration="9.07013562s" podCreationTimestamp="2025-10-14 08:59:58 +0000 UTC" firstStartedPulling="2025-10-14 08:59:59.970750819 +0000 UTC m=+7816.554797446" lastFinishedPulling="2025-10-14 09:00:06.064945888 +0000 UTC m=+7822.648992555" observedRunningTime="2025-10-14 09:00:07.068119163 +0000 UTC m=+7823.652165810" watchObservedRunningTime="2025-10-14 09:00:07.07013562 +0000 UTC m=+7823.654182237" Oct 14 09:00:08 crc kubenswrapper[5018]: I1014 09:00:08.941244 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:08 crc kubenswrapper[5018]: I1014 09:00:08.941284 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:10 crc kubenswrapper[5018]: I1014 09:00:10.002842 5018 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f8l5t" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="registry-server" probeResult="failure" output=< Oct 14 09:00:10 crc kubenswrapper[5018]: timeout: failed to connect service ":50051" within 1s Oct 14 09:00:10 crc kubenswrapper[5018]: > Oct 14 09:00:15 crc kubenswrapper[5018]: I1014 09:00:15.034691 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-527dq"] Oct 14 09:00:15 crc kubenswrapper[5018]: I1014 09:00:15.042327 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-527dq"] Oct 14 09:00:15 crc kubenswrapper[5018]: I1014 09:00:15.692224 5018 scope.go:117] "RemoveContainer" containerID="76ccaf93fe7fdd9fe41953b857e47baabdb375d2fba0a30c8256d88e04add86c" Oct 14 09:00:15 crc kubenswrapper[5018]: I1014 09:00:15.726901 5018 scope.go:117] "RemoveContainer" containerID="ce5d39a39c7ed7bbc9152f97a9bf47f809f0c78524ebd4d01a7d0eea008d0b17" Oct 14 09:00:15 crc kubenswrapper[5018]: I1014 09:00:15.791468 5018 scope.go:117] "RemoveContainer" containerID="62b082682f42b865dbdfcdf0a9e426152caf523ebbdc379a94547fffe07b3870" Oct 14 09:00:16 crc kubenswrapper[5018]: I1014 09:00:16.620857 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11428a7e-0d78-45ed-b1c2-1cbc90fe51cf" path="/var/lib/kubelet/pods/11428a7e-0d78-45ed-b1c2-1cbc90fe51cf/volumes" Oct 14 09:00:19 crc kubenswrapper[5018]: I1014 09:00:19.004881 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:19 crc kubenswrapper[5018]: I1014 09:00:19.058754 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:19 crc kubenswrapper[5018]: I1014 09:00:19.250819 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.225078 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f8l5t" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="registry-server" containerID="cri-o://785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259" gracePeriod=2 Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.609064 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:00:20 crc kubenswrapper[5018]: E1014 09:00:20.609969 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.669971 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.702648 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content\") pod \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.702776 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities\") pod \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.702882 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdjq8\" (UniqueName: \"kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8\") pod \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\" (UID: \"cc8b7d7d-f39b-493f-92ec-e495a46b7b86\") " Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.705742 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities" (OuterVolumeSpecName: "utilities") pod "cc8b7d7d-f39b-493f-92ec-e495a46b7b86" (UID: "cc8b7d7d-f39b-493f-92ec-e495a46b7b86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.715429 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8" (OuterVolumeSpecName: "kube-api-access-qdjq8") pod "cc8b7d7d-f39b-493f-92ec-e495a46b7b86" (UID: "cc8b7d7d-f39b-493f-92ec-e495a46b7b86"). InnerVolumeSpecName "kube-api-access-qdjq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.805994 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.806051 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdjq8\" (UniqueName: \"kubernetes.io/projected/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-kube-api-access-qdjq8\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.820289 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc8b7d7d-f39b-493f-92ec-e495a46b7b86" (UID: "cc8b7d7d-f39b-493f-92ec-e495a46b7b86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:00:20 crc kubenswrapper[5018]: I1014 09:00:20.907341 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8b7d7d-f39b-493f-92ec-e495a46b7b86-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.242405 5018 generic.go:334] "Generic (PLEG): container finished" podID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerID="785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259" exitCode=0 Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.242494 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8l5t" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.242513 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerDied","Data":"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259"} Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.244839 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8l5t" event={"ID":"cc8b7d7d-f39b-493f-92ec-e495a46b7b86","Type":"ContainerDied","Data":"1b96d4330d7e61bd75906732a6f8418c8a375f7e35bfce4bffa6ffe8f0faf151"} Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.245002 5018 scope.go:117] "RemoveContainer" containerID="785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.278136 5018 scope.go:117] "RemoveContainer" containerID="d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.303082 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.316181 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f8l5t"] Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.322400 5018 scope.go:117] "RemoveContainer" containerID="cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.377472 5018 scope.go:117] "RemoveContainer" containerID="785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259" Oct 14 09:00:21 crc kubenswrapper[5018]: E1014 09:00:21.378245 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259\": container with ID starting with 785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259 not found: ID does not exist" containerID="785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.378281 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259"} err="failed to get container status \"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259\": rpc error: code = NotFound desc = could not find container \"785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259\": container with ID starting with 785185361b27ec7a0dbc05e4bb32e5e8210481a9a2604ea33eac339f88f4c259 not found: ID does not exist" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.378305 5018 scope.go:117] "RemoveContainer" containerID="d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335" Oct 14 09:00:21 crc kubenswrapper[5018]: E1014 09:00:21.379032 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335\": container with ID starting with d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335 not found: ID does not exist" containerID="d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.379094 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335"} err="failed to get container status \"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335\": rpc error: code = NotFound desc = could not find container \"d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335\": container with ID starting with d7cee2332ed69fc94cfbfb89ffe02b98eeaabc6d257afa258b67cc9f8b0b3335 not found: ID does not exist" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.379132 5018 scope.go:117] "RemoveContainer" containerID="cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11" Oct 14 09:00:21 crc kubenswrapper[5018]: E1014 09:00:21.379644 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11\": container with ID starting with cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11 not found: ID does not exist" containerID="cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11" Oct 14 09:00:21 crc kubenswrapper[5018]: I1014 09:00:21.379709 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11"} err="failed to get container status \"cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11\": rpc error: code = NotFound desc = could not find container \"cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11\": container with ID starting with cb1127e126ac64d3e0fc19d1d66bbc04ec1aa742899fb6679ae1bf536e3dca11 not found: ID does not exist" Oct 14 09:00:22 crc kubenswrapper[5018]: I1014 09:00:22.641264 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" path="/var/lib/kubelet/pods/cc8b7d7d-f39b-493f-92ec-e495a46b7b86/volumes" Oct 14 09:00:25 crc kubenswrapper[5018]: I1014 09:00:25.047103 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-85f4-account-create-chbwd"] Oct 14 09:00:25 crc kubenswrapper[5018]: I1014 09:00:25.065182 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-85f4-account-create-chbwd"] Oct 14 09:00:26 crc kubenswrapper[5018]: I1014 09:00:26.622971 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="485a9384-0ea8-4df7-922f-caffb6440309" path="/var/lib/kubelet/pods/485a9384-0ea8-4df7-922f-caffb6440309/volumes" Oct 14 09:00:32 crc kubenswrapper[5018]: I1014 09:00:32.608154 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:00:32 crc kubenswrapper[5018]: E1014 09:00:32.609429 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:00:35 crc kubenswrapper[5018]: I1014 09:00:35.078188 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-lk42t"] Oct 14 09:00:35 crc kubenswrapper[5018]: I1014 09:00:35.083201 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-lk42t"] Oct 14 09:00:36 crc kubenswrapper[5018]: I1014 09:00:36.626926 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="871eab4b-2600-4031-a4bd-a8d6e5e1e0bd" path="/var/lib/kubelet/pods/871eab4b-2600-4031-a4bd-a8d6e5e1e0bd/volumes" Oct 14 09:00:44 crc kubenswrapper[5018]: I1014 09:00:44.621810 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:00:44 crc kubenswrapper[5018]: E1014 09:00:44.623081 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:00:57 crc kubenswrapper[5018]: I1014 09:00:57.606281 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:00:57 crc kubenswrapper[5018]: E1014 09:00:57.607380 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.180896 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29340541-h7n4r"] Oct 14 09:01:00 crc kubenswrapper[5018]: E1014 09:01:00.181750 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a68fb7-9603-41e6-a75a-40b35c5c041d" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.181766 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a68fb7-9603-41e6-a75a-40b35c5c041d" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5018]: E1014 09:01:00.181775 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="extract-content" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.181781 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="extract-content" Oct 14 09:01:00 crc kubenswrapper[5018]: E1014 09:01:00.181811 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="registry-server" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.181818 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="registry-server" Oct 14 09:01:00 crc kubenswrapper[5018]: E1014 09:01:00.181839 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="extract-utilities" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.181846 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="extract-utilities" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.182038 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a68fb7-9603-41e6-a75a-40b35c5c041d" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.182053 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8b7d7d-f39b-493f-92ec-e495a46b7b86" containerName="registry-server" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.182977 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.195008 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-h7n4r"] Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.349560 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv7vr\" (UniqueName: \"kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.349717 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.350444 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.350878 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.452827 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.453179 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.453378 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv7vr\" (UniqueName: \"kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.453545 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.465546 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.465816 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.466162 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.487483 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv7vr\" (UniqueName: \"kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr\") pod \"keystone-cron-29340541-h7n4r\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:00 crc kubenswrapper[5018]: I1014 09:01:00.555281 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:01 crc kubenswrapper[5018]: I1014 09:01:01.094600 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-h7n4r"] Oct 14 09:01:01 crc kubenswrapper[5018]: I1014 09:01:01.799441 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-h7n4r" event={"ID":"f117dc78-c8fa-43ce-922c-6c035ff929ac","Type":"ContainerStarted","Data":"c3017755a0e2f72721deb1e16c3269c756b7cd3d3374013490a920e104ae5d49"} Oct 14 09:01:01 crc kubenswrapper[5018]: I1014 09:01:01.799820 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-h7n4r" event={"ID":"f117dc78-c8fa-43ce-922c-6c035ff929ac","Type":"ContainerStarted","Data":"f7a62179cd1d7ba9fbcece377cc29fd914bb901fe62cac230c79a9bc0a42d623"} Oct 14 09:01:01 crc kubenswrapper[5018]: I1014 09:01:01.833249 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29340541-h7n4r" podStartSLOduration=1.833226708 podStartE2EDuration="1.833226708s" podCreationTimestamp="2025-10-14 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:01.828888345 +0000 UTC m=+7878.412935002" watchObservedRunningTime="2025-10-14 09:01:01.833226708 +0000 UTC m=+7878.417273355" Oct 14 09:01:03 crc kubenswrapper[5018]: I1014 09:01:03.826818 5018 generic.go:334] "Generic (PLEG): container finished" podID="f117dc78-c8fa-43ce-922c-6c035ff929ac" containerID="c3017755a0e2f72721deb1e16c3269c756b7cd3d3374013490a920e104ae5d49" exitCode=0 Oct 14 09:01:03 crc kubenswrapper[5018]: I1014 09:01:03.826898 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-h7n4r" event={"ID":"f117dc78-c8fa-43ce-922c-6c035ff929ac","Type":"ContainerDied","Data":"c3017755a0e2f72721deb1e16c3269c756b7cd3d3374013490a920e104ae5d49"} Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.314301 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.489381 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data\") pod \"f117dc78-c8fa-43ce-922c-6c035ff929ac\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.489556 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv7vr\" (UniqueName: \"kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr\") pod \"f117dc78-c8fa-43ce-922c-6c035ff929ac\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.489651 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys\") pod \"f117dc78-c8fa-43ce-922c-6c035ff929ac\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.489707 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle\") pod \"f117dc78-c8fa-43ce-922c-6c035ff929ac\" (UID: \"f117dc78-c8fa-43ce-922c-6c035ff929ac\") " Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.494552 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f117dc78-c8fa-43ce-922c-6c035ff929ac" (UID: "f117dc78-c8fa-43ce-922c-6c035ff929ac"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.496222 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr" (OuterVolumeSpecName: "kube-api-access-lv7vr") pod "f117dc78-c8fa-43ce-922c-6c035ff929ac" (UID: "f117dc78-c8fa-43ce-922c-6c035ff929ac"). InnerVolumeSpecName "kube-api-access-lv7vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.516959 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f117dc78-c8fa-43ce-922c-6c035ff929ac" (UID: "f117dc78-c8fa-43ce-922c-6c035ff929ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.555109 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data" (OuterVolumeSpecName: "config-data") pod "f117dc78-c8fa-43ce-922c-6c035ff929ac" (UID: "f117dc78-c8fa-43ce-922c-6c035ff929ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.592073 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.592107 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv7vr\" (UniqueName: \"kubernetes.io/projected/f117dc78-c8fa-43ce-922c-6c035ff929ac-kube-api-access-lv7vr\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.592118 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.592128 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f117dc78-c8fa-43ce-922c-6c035ff929ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.857933 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-h7n4r" event={"ID":"f117dc78-c8fa-43ce-922c-6c035ff929ac","Type":"ContainerDied","Data":"f7a62179cd1d7ba9fbcece377cc29fd914bb901fe62cac230c79a9bc0a42d623"} Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.858264 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7a62179cd1d7ba9fbcece377cc29fd914bb901fe62cac230c79a9bc0a42d623" Oct 14 09:01:05 crc kubenswrapper[5018]: I1014 09:01:05.858033 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-h7n4r" Oct 14 09:01:09 crc kubenswrapper[5018]: I1014 09:01:09.607204 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:01:09 crc kubenswrapper[5018]: E1014 09:01:09.608352 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:01:15 crc kubenswrapper[5018]: I1014 09:01:15.931282 5018 scope.go:117] "RemoveContainer" containerID="3511b3411f0b272e1d3c9f02eb319b73fb65032a2dda5b0a0d85238a7ca5bdcd" Oct 14 09:01:15 crc kubenswrapper[5018]: I1014 09:01:15.974395 5018 scope.go:117] "RemoveContainer" containerID="5391d3a26510d78f54511f3524ac18659f17e980c997ac94cdfe99f6c3216466" Oct 14 09:01:16 crc kubenswrapper[5018]: I1014 09:01:16.049596 5018 scope.go:117] "RemoveContainer" containerID="0474eb461f962bcb1871b0fbba81d351ab2eccb58550fc06be029fe0f3ec1316" Oct 14 09:01:22 crc kubenswrapper[5018]: I1014 09:01:22.605886 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:01:22 crc kubenswrapper[5018]: E1014 09:01:22.606957 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:01:35 crc kubenswrapper[5018]: I1014 09:01:35.604809 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:01:35 crc kubenswrapper[5018]: E1014 09:01:35.606076 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:01:37 crc kubenswrapper[5018]: I1014 09:01:37.061040 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-p2mrn"] Oct 14 09:01:37 crc kubenswrapper[5018]: I1014 09:01:37.074420 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-m8j5s"] Oct 14 09:01:37 crc kubenswrapper[5018]: I1014 09:01:37.082107 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-p2mrn"] Oct 14 09:01:37 crc kubenswrapper[5018]: I1014 09:01:37.089035 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-m8j5s"] Oct 14 09:01:38 crc kubenswrapper[5018]: I1014 09:01:38.031715 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-qrvbw"] Oct 14 09:01:38 crc kubenswrapper[5018]: I1014 09:01:38.039149 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-qrvbw"] Oct 14 09:01:38 crc kubenswrapper[5018]: I1014 09:01:38.616862 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab5c795-ddf6-4f81-bd9e-f15389bbba38" path="/var/lib/kubelet/pods/0ab5c795-ddf6-4f81-bd9e-f15389bbba38/volumes" Oct 14 09:01:38 crc kubenswrapper[5018]: I1014 09:01:38.617714 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6494b4a3-ba5b-4eda-9825-7f5b3acefd5f" path="/var/lib/kubelet/pods/6494b4a3-ba5b-4eda-9825-7f5b3acefd5f/volumes" Oct 14 09:01:38 crc kubenswrapper[5018]: I1014 09:01:38.618182 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b05444e-ffec-436a-bb1f-1c1b31ebf7a8" path="/var/lib/kubelet/pods/7b05444e-ffec-436a-bb1f-1c1b31ebf7a8/volumes" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.244650 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:01:41 crc kubenswrapper[5018]: E1014 09:01:41.245396 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f117dc78-c8fa-43ce-922c-6c035ff929ac" containerName="keystone-cron" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.245413 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f117dc78-c8fa-43ce-922c-6c035ff929ac" containerName="keystone-cron" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.245915 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f117dc78-c8fa-43ce-922c-6c035ff929ac" containerName="keystone-cron" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.247699 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.261227 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.359548 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.359845 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.359943 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bbk8\" (UniqueName: \"kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.461960 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.462087 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.462126 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bbk8\" (UniqueName: \"kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.462944 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.463320 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.483041 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bbk8\" (UniqueName: \"kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8\") pod \"certified-operators-l57bp\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:41 crc kubenswrapper[5018]: I1014 09:01:41.585373 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:42 crc kubenswrapper[5018]: I1014 09:01:42.120150 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:01:42 crc kubenswrapper[5018]: I1014 09:01:42.318398 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerStarted","Data":"6eec88e8f42e457873b962117a6142f6b81ba76b59599c665a244fc4632fc947"} Oct 14 09:01:43 crc kubenswrapper[5018]: I1014 09:01:43.332756 5018 generic.go:334] "Generic (PLEG): container finished" podID="78c654d8-695b-4f27-be10-8e5861f7645d" containerID="0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e" exitCode=0 Oct 14 09:01:43 crc kubenswrapper[5018]: I1014 09:01:43.333399 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerDied","Data":"0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e"} Oct 14 09:01:45 crc kubenswrapper[5018]: I1014 09:01:45.360390 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerStarted","Data":"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772"} Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.047143 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-aea5-account-create-gfrn8"] Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.067215 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-aea5-account-create-gfrn8"] Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.377610 5018 generic.go:334] "Generic (PLEG): container finished" podID="78c654d8-695b-4f27-be10-8e5861f7645d" containerID="ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772" exitCode=0 Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.377805 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerDied","Data":"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772"} Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.606340 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:01:46 crc kubenswrapper[5018]: E1014 09:01:46.606725 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:01:46 crc kubenswrapper[5018]: I1014 09:01:46.624211 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c7cdc3-830e-416d-9218-561f257c524a" path="/var/lib/kubelet/pods/d5c7cdc3-830e-416d-9218-561f257c524a/volumes" Oct 14 09:01:47 crc kubenswrapper[5018]: I1014 09:01:47.041358 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8298-account-create-jtn5n"] Oct 14 09:01:47 crc kubenswrapper[5018]: I1014 09:01:47.058320 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8298-account-create-jtn5n"] Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.057205 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-eeba-account-create-mfhsn"] Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.067383 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-eeba-account-create-mfhsn"] Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.406231 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerStarted","Data":"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25"} Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.443912 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l57bp" podStartSLOduration=2.738850064 podStartE2EDuration="7.44388638s" podCreationTimestamp="2025-10-14 09:01:41 +0000 UTC" firstStartedPulling="2025-10-14 09:01:43.33656288 +0000 UTC m=+7919.920609507" lastFinishedPulling="2025-10-14 09:01:48.041599196 +0000 UTC m=+7924.625645823" observedRunningTime="2025-10-14 09:01:48.432354863 +0000 UTC m=+7925.016401530" watchObservedRunningTime="2025-10-14 09:01:48.44388638 +0000 UTC m=+7925.027933047" Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.619517 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f3934a-36ce-494e-8c56-707748aaf954" path="/var/lib/kubelet/pods/02f3934a-36ce-494e-8c56-707748aaf954/volumes" Oct 14 09:01:48 crc kubenswrapper[5018]: I1014 09:01:48.620476 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9461fc57-4388-43b1-84d1-08c4231f8aad" path="/var/lib/kubelet/pods/9461fc57-4388-43b1-84d1-08c4231f8aad/volumes" Oct 14 09:01:51 crc kubenswrapper[5018]: I1014 09:01:51.586091 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:51 crc kubenswrapper[5018]: I1014 09:01:51.586583 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:01:51 crc kubenswrapper[5018]: I1014 09:01:51.676916 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:02:00 crc kubenswrapper[5018]: I1014 09:02:00.607537 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:02:00 crc kubenswrapper[5018]: E1014 09:02:00.609237 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:02:01 crc kubenswrapper[5018]: I1014 09:02:01.658569 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:02:01 crc kubenswrapper[5018]: I1014 09:02:01.725132 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:02:02 crc kubenswrapper[5018]: I1014 09:02:02.602596 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l57bp" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="registry-server" containerID="cri-o://2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25" gracePeriod=2 Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.140514 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.203160 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities\") pod \"78c654d8-695b-4f27-be10-8e5861f7645d\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.203312 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content\") pod \"78c654d8-695b-4f27-be10-8e5861f7645d\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.203530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bbk8\" (UniqueName: \"kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8\") pod \"78c654d8-695b-4f27-be10-8e5861f7645d\" (UID: \"78c654d8-695b-4f27-be10-8e5861f7645d\") " Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.204003 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities" (OuterVolumeSpecName: "utilities") pod "78c654d8-695b-4f27-be10-8e5861f7645d" (UID: "78c654d8-695b-4f27-be10-8e5861f7645d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.204312 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.210756 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8" (OuterVolumeSpecName: "kube-api-access-7bbk8") pod "78c654d8-695b-4f27-be10-8e5861f7645d" (UID: "78c654d8-695b-4f27-be10-8e5861f7645d"). InnerVolumeSpecName "kube-api-access-7bbk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.272997 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78c654d8-695b-4f27-be10-8e5861f7645d" (UID: "78c654d8-695b-4f27-be10-8e5861f7645d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.306715 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78c654d8-695b-4f27-be10-8e5861f7645d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.306771 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bbk8\" (UniqueName: \"kubernetes.io/projected/78c654d8-695b-4f27-be10-8e5861f7645d-kube-api-access-7bbk8\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.617717 5018 generic.go:334] "Generic (PLEG): container finished" podID="78c654d8-695b-4f27-be10-8e5861f7645d" containerID="2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25" exitCode=0 Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.617980 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerDied","Data":"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25"} Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.618089 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l57bp" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.618197 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l57bp" event={"ID":"78c654d8-695b-4f27-be10-8e5861f7645d","Type":"ContainerDied","Data":"6eec88e8f42e457873b962117a6142f6b81ba76b59599c665a244fc4632fc947"} Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.618247 5018 scope.go:117] "RemoveContainer" containerID="2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.641364 5018 scope.go:117] "RemoveContainer" containerID="ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.666568 5018 scope.go:117] "RemoveContainer" containerID="0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.672495 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.683153 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l57bp"] Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.721005 5018 scope.go:117] "RemoveContainer" containerID="2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25" Oct 14 09:02:03 crc kubenswrapper[5018]: E1014 09:02:03.721804 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25\": container with ID starting with 2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25 not found: ID does not exist" containerID="2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.721857 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25"} err="failed to get container status \"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25\": rpc error: code = NotFound desc = could not find container \"2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25\": container with ID starting with 2c1ac57fc7de6015b51a4754b3c8e1edd6eea5822dce7ca58c003ffc836a0b25 not found: ID does not exist" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.721890 5018 scope.go:117] "RemoveContainer" containerID="ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772" Oct 14 09:02:03 crc kubenswrapper[5018]: E1014 09:02:03.722284 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772\": container with ID starting with ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772 not found: ID does not exist" containerID="ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.722313 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772"} err="failed to get container status \"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772\": rpc error: code = NotFound desc = could not find container \"ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772\": container with ID starting with ec7b2835783932eec7fc3c3f2340d8ecd2a47f636669452f86e4d5b2b61f7772 not found: ID does not exist" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.722332 5018 scope.go:117] "RemoveContainer" containerID="0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e" Oct 14 09:02:03 crc kubenswrapper[5018]: E1014 09:02:03.722645 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e\": container with ID starting with 0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e not found: ID does not exist" containerID="0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e" Oct 14 09:02:03 crc kubenswrapper[5018]: I1014 09:02:03.722805 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e"} err="failed to get container status \"0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e\": rpc error: code = NotFound desc = could not find container \"0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e\": container with ID starting with 0946d8b4ce3f89726e649132755f4e57b72a04b079fb2985ee6dac0982a8cb9e not found: ID does not exist" Oct 14 09:02:04 crc kubenswrapper[5018]: I1014 09:02:04.619143 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" path="/var/lib/kubelet/pods/78c654d8-695b-4f27-be10-8e5861f7645d/volumes" Oct 14 09:02:05 crc kubenswrapper[5018]: I1014 09:02:05.060704 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdg8x"] Oct 14 09:02:05 crc kubenswrapper[5018]: I1014 09:02:05.071367 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdg8x"] Oct 14 09:02:06 crc kubenswrapper[5018]: I1014 09:02:06.616373 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c" path="/var/lib/kubelet/pods/44c7e1f8-4d12-4732-8eb9-d99a6bbd9d8c/volumes" Oct 14 09:02:12 crc kubenswrapper[5018]: I1014 09:02:12.605274 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:02:12 crc kubenswrapper[5018]: E1014 09:02:12.606564 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.226540 5018 scope.go:117] "RemoveContainer" containerID="b524bc97d5ef2425dbc03e111632232af0f5a589133c84eec564c80c53abb4af" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.262480 5018 scope.go:117] "RemoveContainer" containerID="bd1e1e82e5ba7088d371b42b80db29cf7e78000b1ebbcd9beb7cafe27869faa3" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.324002 5018 scope.go:117] "RemoveContainer" containerID="3b4139a5719d9a5784da94eef162c449f1bc1776b8058820ae822f1d0f8f4ce1" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.376316 5018 scope.go:117] "RemoveContainer" containerID="1e0ab488798d47de0be9afd2be9d4121397609cc00e13ec274b3fa8cab4cf8c7" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.423065 5018 scope.go:117] "RemoveContainer" containerID="bf5e3c508a939f648024a637db2a5b33ea30c5568dc14e1a60d8cc63b19e066c" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.466302 5018 scope.go:117] "RemoveContainer" containerID="cdb384054bfc015a9477cd485c9b6d9ebbb34e899544fa170bed94271ce4a0ea" Oct 14 09:02:16 crc kubenswrapper[5018]: I1014 09:02:16.537006 5018 scope.go:117] "RemoveContainer" containerID="694bc6842c5310eb5784443688aa5e6a81785acbd4e872a7bfe75437a6415cb7" Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.560767 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.561414 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" containerName="openstackclient" containerID="cri-o://ae4fa989c0d49cc43c46cfbb1bf94f22382f8020c2ea5b17401082ece4256593" gracePeriod=2 Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.615822 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.647200 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.647426 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-log" containerID="cri-o://ba05b131d9cc289cc0f7d05bb2b92055b3494f32ba43fa2269cb65376bf6bbb9" gracePeriod=30 Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.650993 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-httpd" containerID="cri-o://d43f95354a677e6c3aafe9f165dbe59e283b3bddc5f2df6c04584560150ca6b8" gracePeriod=30 Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.829917 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.925079 5018 generic.go:334] "Generic (PLEG): container finished" podID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerID="ba05b131d9cc289cc0f7d05bb2b92055b3494f32ba43fa2269cb65376bf6bbb9" exitCode=143 Oct 14 09:02:19 crc kubenswrapper[5018]: I1014 09:02:19.925386 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerDied","Data":"ba05b131d9cc289cc0f7d05bb2b92055b3494f32ba43fa2269cb65376bf6bbb9"} Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.004796 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.004870 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:20.50485239 +0000 UTC m=+7957.088899017 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data") pod "rabbitmq-cell1-server-0" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.032494 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5b96cbd67f-h4t2t"] Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.032904 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" containerName="openstackclient" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.032919 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" containerName="openstackclient" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.032930 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="extract-utilities" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.032937 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="extract-utilities" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.032960 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="extract-content" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.032967 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="extract-content" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.032999 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="registry-server" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.033005 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="registry-server" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.033189 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c654d8-695b-4f27-be10-8e5861f7645d" containerName="registry-server" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.033214 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" containerName="openstackclient" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.034030 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.080817 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5b96cbd67f-h4t2t"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.142169 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.142414 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.142475 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.142506 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.178530 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-59d586bc68-bvmll"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.196227 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.265819 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-59d586bc68-bvmll"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.270977 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271139 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271249 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271326 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271428 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271540 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271610 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271703 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.271935 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.272043 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.272297 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.272404 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:20.772388074 +0000 UTC m=+7957.356434701 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.295504 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.312381 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.323920 5018 projected.go:194] Error preparing data for projected volume kube-api-access-sdcgs for pod openstack/heat-engine-5b96cbd67f-h4t2t: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.324164 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:20.824147065 +0000 UTC m=+7957.408193692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sdcgs" (UniqueName: "kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.354579 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.356093 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374525 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374628 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374670 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374692 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374711 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374818 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.374864 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrcz4\" (UniqueName: \"kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4\") pod \"neutrond3e5-account-delete-4mrl8\" (UID: \"eab7a625-e4b8-4038-b9a7-ce1f7af786bf\") " pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.375217 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.375261 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:20.875248188 +0000 UTC m=+7957.459294815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.379692 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.390343 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.390930 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.391342 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.391529 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.402673 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.441787 5018 projected.go:194] Error preparing data for projected volume kube-api-access-lhqcc for pod openstack/heat-api-59d586bc68-bvmll: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.441860 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:20.94183953 +0000 UTC m=+7957.525886157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-lhqcc" (UniqueName: "kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.453359 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.454893 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.475881 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.476260 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-1" podUID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerName="openstack-network-exporter" containerID="cri-o://f5a9062ecbff01d430f8c45450a3ea8dfae4ad161b221c5afda3f75ccde512b0" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.482095 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrcz4\" (UniqueName: \"kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4\") pod \"neutrond3e5-account-delete-4mrl8\" (UID: \"eab7a625-e4b8-4038-b9a7-ce1f7af786bf\") " pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.501680 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.502058 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="openstack-network-exporter" containerID="cri-o://b8af3ead5cf9c4d299de3975f6ab77037e7e1da24c12c876f67ca25674008762" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.541181 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.541528 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-2" podUID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerName="openstack-network-exporter" containerID="cri-o://19de2087aaf37731d242d90edcb3f4e4a539c73a836422704d5ea34ead7356dc" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.574264 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrcz4\" (UniqueName: \"kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4\") pod \"neutrond3e5-account-delete-4mrl8\" (UID: \"eab7a625-e4b8-4038-b9a7-ce1f7af786bf\") " pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.599350 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6rwq\" (UniqueName: \"kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq\") pod \"barbican3632-account-delete-sgfbk\" (UID: \"8fb6f85b-5866-46bf-a826-0a22243318c1\") " pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.599583 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.604579 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.605198 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.615890 5018 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/horizon-74b99bc44-qsr8p" secret="" err="secret \"horizon-horizon-dockercfg-sw76v\" not found" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.622718 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data podName:29481fb2-99c9-41f0-b650-a971fa5ff28b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.099630285 +0000 UTC m=+7957.683676912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data") pod "rabbitmq-server-0" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b") : configmap "rabbitmq-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.622815 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.622775283 +0000 UTC m=+7958.206821910 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data") pod "rabbitmq-cell1-server-0" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.668926 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="ovsdbserver-sb" containerID="cri-o://fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.670483 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.670691 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" containerID="cri-o://d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" gracePeriod=30 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.670976 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="openstack-network-exporter" containerID="cri-o://0cf0d1c6030e09c2eda77accb9e964aeb0ff1f0dbf949d67d6d6824cf055ea0a" gracePeriod=30 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.692679 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.694038 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.703869 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6rwq\" (UniqueName: \"kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq\") pod \"barbican3632-account-delete-sgfbk\" (UID: \"8fb6f85b-5866-46bf-a826-0a22243318c1\") " pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705340 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705631 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705669 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.205655039 +0000 UTC m=+7957.789701666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705694 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.20568816 +0000 UTC m=+7957.789734787 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705826 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.705879 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.205863875 +0000 UTC m=+7957.789910502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.759276 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6rwq\" (UniqueName: \"kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq\") pod \"barbican3632-account-delete-sgfbk\" (UID: \"8fb6f85b-5866-46bf-a826-0a22243318c1\") " pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.766773 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.778075 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.796540 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.797543 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-1" podUID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerName="openstack-network-exporter" containerID="cri-o://f0b29be9a0afeb5e7192333db87e868120fdaf92944e86c96054f44887940f2f" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.828894 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.829463 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="openstack-network-exporter" containerID="cri-o://e06750d462802f2f98ea3875ae3f637ed5b88aae6f70ae700e5336271d04d93c" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.838298 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.838395 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.838527 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k2m7\" (UniqueName: \"kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7\") pod \"heatd90b-account-delete-sh6hm\" (UID: \"90c10727-d872-4046-876c-39475c9d8623\") " pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.838825 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.838909 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.838889556 +0000 UTC m=+7958.422936183 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.850309 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.868312 5018 projected.go:194] Error preparing data for projected volume kube-api-access-sdcgs for pod openstack/heat-engine-5b96cbd67f-h4t2t: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.868374 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.868357774 +0000 UTC m=+7958.452404401 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-sdcgs" (UniqueName: "kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.868749 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-2" podUID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerName="openstack-network-exporter" containerID="cri-o://8518cc00808c52898e754e85cacd8523a46b73a34084d59361518ea964572758" gracePeriod=300 Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.869217 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.922146 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.924479 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.940043 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k2m7\" (UniqueName: \"kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7\") pod \"heatd90b-account-delete-sh6hm\" (UID: \"90c10727-d872-4046-876c-39475c9d8623\") " pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.940173 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.941101 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: E1014 09:02:20.941165 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:21.941146112 +0000 UTC m=+7958.525192739 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : secret "heat-config-data" not found Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.944892 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.974190 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k2m7\" (UniqueName: \"kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7\") pod \"heatd90b-account-delete-sh6hm\" (UID: \"90c10727-d872-4046-876c-39475c9d8623\") " pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.975091 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.976880 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:20 crc kubenswrapper[5018]: I1014 09:02:20.998650 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.011664 5018 generic.go:334] "Generic (PLEG): container finished" podID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerID="f5a9062ecbff01d430f8c45450a3ea8dfae4ad161b221c5afda3f75ccde512b0" exitCode=2 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.011709 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerDied","Data":"f5a9062ecbff01d430f8c45450a3ea8dfae4ad161b221c5afda3f75ccde512b0"} Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.043580 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfcb8\" (UniqueName: \"kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8\") pod \"glance72c6-account-delete-27sjg\" (UID: \"7f55eb6e-96bb-4019-83a9-10ba9f455ecc\") " pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.043887 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hffn\" (UniqueName: \"kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn\") pod \"placement85f4-account-delete-lx2cs\" (UID: \"5876dcf2-c564-45ac-bf38-bebf49e5661f\") " pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.044057 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.058361 5018 projected.go:194] Error preparing data for projected volume kube-api-access-lhqcc for pod openstack/heat-api-59d586bc68-bvmll: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.058473 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.058441516 +0000 UTC m=+7958.642488143 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-lhqcc" (UniqueName: "kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.061287 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.125437 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-6ssgc"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.147794 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hffn\" (UniqueName: \"kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn\") pod \"placement85f4-account-delete-lx2cs\" (UID: \"5876dcf2-c564-45ac-bf38-bebf49e5661f\") " pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.148202 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfcb8\" (UniqueName: \"kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8\") pod \"glance72c6-account-delete-27sjg\" (UID: \"7f55eb6e-96bb-4019-83a9-10ba9f455ecc\") " pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.148658 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.148698 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data podName:29481fb2-99c9-41f0-b650-a971fa5ff28b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.148684291 +0000 UTC m=+7958.732730918 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data") pod "rabbitmq-server-0" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b") : configmap "rabbitmq-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.159262 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-6ssgc"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.192270 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfcb8\" (UniqueName: \"kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8\") pod \"glance72c6-account-delete-27sjg\" (UID: \"7f55eb6e-96bb-4019-83a9-10ba9f455ecc\") " pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.194149 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hffn\" (UniqueName: \"kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn\") pod \"placement85f4-account-delete-lx2cs\" (UID: \"5876dcf2-c564-45ac-bf38-bebf49e5661f\") " pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.228134 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.246600 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255126 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255196 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.255181219 +0000 UTC m=+7958.839227846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255533 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255558 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.255550839 +0000 UTC m=+7958.839597466 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255611 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.255649 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.255642832 +0000 UTC m=+7958.839689449 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.281507 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="ovsdbserver-nb" containerID="cri-o://f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" gracePeriod=300 Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.369728 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.391040 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.392275 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-52cnz"] Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.394284 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.394453 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.418044 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.420215 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74b99bc44-qsr8p" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" containerID="cri-o://392ee2ce4169eab5c3bfa508967090781008c6bb4b91e91bc63269bdd41562f3" gracePeriod=30 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.418733 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74b99bc44-qsr8p" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon-log" containerID="cri-o://dcb994f2a796078d866b79a0686477e41a0d979852b18ac2243d743ecb903f27" gracePeriod=30 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.438978 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-52cnz"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.589689 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.592897 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="dnsmasq-dns" containerID="cri-o://5045a096ebb1e46ec923b34da32996ecc7663e4e86bfba2ce6bb474b4f341eec" gracePeriod=10 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.605660 5018 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican-worker-854795b57-c4dnm" secret="" err="secret \"barbican-barbican-dockercfg-tljlw\" not found" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.678066 5018 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.678147 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.17813162 +0000 UTC m=+7958.762178247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-worker-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.678418 5018 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.678462 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:22.178452899 +0000 UTC m=+7958.762499526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.680042 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.680110 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:23.680076966 +0000 UTC m=+7960.264123593 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data") pod "rabbitmq-cell1-server-0" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.913639 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a is running failed: container process not found" containerID="f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.915332 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.915675 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69dd4dcdf-tjqrh" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-api" containerID="cri-o://645fd93c4abe65ac53978d4f7400e542a575a47bf0b11107c5f6b7d78f3348dd" gracePeriod=30 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.916214 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69dd4dcdf-tjqrh" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-httpd" containerID="cri-o://b153baeb0241dacb32e954510d08cddc263a2d3ee9e254846604179d4e462655" gracePeriod=30 Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.917508 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.917599 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.917862 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.917938 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:23.917911396 +0000 UTC m=+7960.501958023 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : secret "heat-config-data" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.922993 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a is running failed: container process not found" containerID="f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.924905 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a is running failed: container process not found" containerID="f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.924938 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="ovsdbserver-nb" Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.930381 5018 projected.go:194] Error preparing data for projected volume kube-api-access-sdcgs for pod openstack/heat-engine-5b96cbd67f-h4t2t: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:21 crc kubenswrapper[5018]: E1014 09:02:21.930451 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:23.930434802 +0000 UTC m=+7960.514481429 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-sdcgs" (UniqueName: "kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.987495 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:21 crc kubenswrapper[5018]: I1014 09:02:21.991485 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.022150 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.022277 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbkmx\" (UniqueName: \"kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx\") pod \"aodh033c-account-delete-9nfhv\" (UID: \"28d51170-15aa-4edd-9223-ebe057d805b0\") " pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.022480 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.022542 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.022523429 +0000 UTC m=+7960.606570056 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : secret "heat-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.036952 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.090368 5018 generic.go:334] "Generic (PLEG): container finished" podID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerID="f0b29be9a0afeb5e7192333db87e868120fdaf92944e86c96054f44887940f2f" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.090450 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerDied","Data":"f0b29be9a0afeb5e7192333db87e868120fdaf92944e86c96054f44887940f2f"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.116021 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ch8tx"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.126198 5018 generic.go:334] "Generic (PLEG): container finished" podID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerID="8518cc00808c52898e754e85cacd8523a46b73a34084d59361518ea964572758" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.126270 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.126288 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerDied","Data":"8518cc00808c52898e754e85cacd8523a46b73a34084d59361518ea964572758"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.126325 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbkmx\" (UniqueName: \"kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx\") pod \"aodh033c-account-delete-9nfhv\" (UID: \"28d51170-15aa-4edd-9223-ebe057d805b0\") " pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.134463 5018 projected.go:194] Error preparing data for projected volume kube-api-access-lhqcc for pod openstack/heat-api-59d586bc68-bvmll: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.134542 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.134522003 +0000 UTC m=+7960.718568640 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-lhqcc" (UniqueName: "kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.140152 5018 generic.go:334] "Generic (PLEG): container finished" podID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerID="0cf0d1c6030e09c2eda77accb9e964aeb0ff1f0dbf949d67d6d6824cf055ea0a" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.140215 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerDied","Data":"0cf0d1c6030e09c2eda77accb9e964aeb0ff1f0dbf949d67d6d6824cf055ea0a"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.152588 5018 generic.go:334] "Generic (PLEG): container finished" podID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerID="19de2087aaf37731d242d90edcb3f4e4a539c73a836422704d5ea34ead7356dc" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.152839 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerDied","Data":"19de2087aaf37731d242d90edcb3f4e4a539c73a836422704d5ea34ead7356dc"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.156488 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbkmx\" (UniqueName: \"kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx\") pod \"aodh033c-account-delete-9nfhv\" (UID: \"28d51170-15aa-4edd-9223-ebe057d805b0\") " pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.167350 5018 generic.go:334] "Generic (PLEG): container finished" podID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerID="5045a096ebb1e46ec923b34da32996ecc7663e4e86bfba2ce6bb474b4f341eec" exitCode=0 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.167410 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" event={"ID":"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e","Type":"ContainerDied","Data":"5045a096ebb1e46ec923b34da32996ecc7663e4e86bfba2ce6bb474b4f341eec"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.168632 5018 generic.go:334] "Generic (PLEG): container finished" podID="e216b65d-5bc3-453c-aa74-b9792e47575e" containerID="ae4fa989c0d49cc43c46cfbb1bf94f22382f8020c2ea5b17401082ece4256593" exitCode=137 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.173782 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-xxvbb"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.180138 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9703a145-2c14-4a43-a767-e6def8fe6063/ovsdbserver-sb/0.log" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.180184 5018 generic.go:334] "Generic (PLEG): container finished" podID="9703a145-2c14-4a43-a767-e6def8fe6063" containerID="b8af3ead5cf9c4d299de3975f6ab77037e7e1da24c12c876f67ca25674008762" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.180201 5018 generic.go:334] "Generic (PLEG): container finished" podID="9703a145-2c14-4a43-a767-e6def8fe6063" containerID="fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" exitCode=143 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.180258 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerDied","Data":"b8af3ead5cf9c4d299de3975f6ab77037e7e1da24c12c876f67ca25674008762"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.180284 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerDied","Data":"fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.181906 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b67bc93c-1c02-47fa-b2a5-57560256e89b/ovsdbserver-nb/0.log" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.181936 5018 generic.go:334] "Generic (PLEG): container finished" podID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerID="e06750d462802f2f98ea3875ae3f637ed5b88aae6f70ae700e5336271d04d93c" exitCode=2 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.181946 5018 generic.go:334] "Generic (PLEG): container finished" podID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerID="f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" exitCode=143 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.181969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerDied","Data":"e06750d462802f2f98ea3875ae3f637ed5b88aae6f70ae700e5336271d04d93c"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.181984 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerDied","Data":"f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a"} Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.216578 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ch8tx"] Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.227827 5018 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.227889 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:23.227873486 +0000 UTC m=+7959.811920113 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-worker-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.228213 5018 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.228237 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:23.228229616 +0000 UTC m=+7959.812276243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.228265 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.228283 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data podName:29481fb2-99c9-41f0-b650-a971fa5ff28b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.228277078 +0000 UTC m=+7960.812323705 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data") pod "rabbitmq-server-0" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b") : configmap "rabbitmq-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.238766 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-xxvbb"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.249875 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.297910 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="rabbitmq" containerID="cri-o://353c1ff5f1490343dd0dfdf4922745f628e70fdbb923936bb2cb5ed954c5db54" gracePeriod=604800 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.301720 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.301985 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-log" containerID="cri-o://b10bff320cd1aef3f540117705e3e341d22490e36c13290cfba7eec50e8f6a71" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.302422 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-httpd" containerID="cri-o://89c2c3c1c29bf6244a053cff8a3b895dd672b6f514f9287e56db1eeba83bd95e" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.312324 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.312851 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5b8df9c49b-fpwlm" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-log" containerID="cri-o://9b40efcd93a346933415044024e90ede001495fcfd8fb0320885bab4c64fda7d" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.313228 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5b8df9c49b-fpwlm" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-api" containerID="cri-o://40fb8b68d8bb04c1d344475d4f00e45548e0b4c28983fcb23f3b012d39221907" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333215 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333282 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.333265162 +0000 UTC m=+7960.917311789 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333336 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333372 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.333364475 +0000 UTC m=+7960.917411102 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333429 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.333449 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.333443587 +0000 UTC m=+7960.917490214 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.353086 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.399227 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.399533 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="cinder-scheduler" containerID="cri-o://3c0372fbfcb7d77786b28550cb757bcf2d36775c128fc8d61c2eeed2bb4a27a4" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.399967 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="probe" containerID="cri-o://ac50df93bbc18e55844ca557b4aa55078662999d23f67f0d58048a5477b2f20c" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.418969 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.429504 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.434140 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-559f8b448b-zgvh5" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-server" containerID="cri-o://f3bfacf0b50c0a9351a97df8cd9a4d4d20587d8f3b2250add56a2512071e49f6" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.434140 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-559f8b448b-zgvh5" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-httpd" containerID="cri-o://8ea98fcf3c27a68c727221403d77e0a9e5e402b4f567534f6bb06689161fc3eb" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.442396 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.442688 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api-log" containerID="cri-o://ce612af3722b887703ef66d47d90f1bdbe0b44f623bc74647fe5ba7231238eed" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.442802 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api" containerID="cri-o://cc378e4ea02d43f477ec8203ef137c674e7055a3b57a90d0bac877f9fdd800da" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.451101 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.506567 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-s48hc"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.540979 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-s48hc"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.599890 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.600167 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-api" containerID="cri-o://2d821f28ca619425ae5555a780634062a928277adbe2e77a11f5825e3ed63b9d" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.600562 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-listener" containerID="cri-o://84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.600608 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-notifier" containerID="cri-o://4d5980364138e871981e0c2a309c64682bdee2ba69842dd2a7eb988420399d1e" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.600662 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-evaluator" containerID="cri-o://a7585a2f1c37963c5dbf63572362ad5333d952de5f762503637a079285e5bd70" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.637478 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="rabbitmq" containerID="cri-o://efb355b1080edb66f752f3f04a2f8a2c9ade67592342b2f5278487f5bddf2b1a" gracePeriod=604800 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.644143 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208ccbe0-ffd5-478f-aca6-8af7d38f9b33" path="/var/lib/kubelet/pods/208ccbe0-ffd5-478f-aca6-8af7d38f9b33/volumes" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.645025 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd8ec80-6c1d-4893-8d7c-765d6966d9ec" path="/var/lib/kubelet/pods/3cd8ec80-6c1d-4893-8d7c-765d6966d9ec/volumes" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.645678 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d71d68-03a3-46e4-98be-ebd0904194d2" path="/var/lib/kubelet/pods/68d71d68-03a3-46e4-98be-ebd0904194d2/volumes" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.664858 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b07a1347-0c41-470f-b4ce-d916787935c0" path="/var/lib/kubelet/pods/b07a1347-0c41-470f-b4ce-d916787935c0/volumes" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.668043 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2" path="/var/lib/kubelet/pods/e5bb1ac7-d8ca-421d-a8f4-bbe605f0d6a2/volumes" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.670798 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.670829 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.670841 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.670853 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.672306 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener-log" containerID="cri-o://0e6d9ef3c618844329be04c22cf07c96416cade5d7551e947ded7ff89d6b027d" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.672707 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener" containerID="cri-o://f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.672879 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-854795b57-c4dnm" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker" containerID="cri-o://74d5736c20baee1bfbfb8ea1b988e45aea425c479ba0be5662af48c4044d6a13" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.671221 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-854795b57-c4dnm" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker-log" containerID="cri-o://e8b14272735d2836638396ff83b7ead777ae1c594b519c1e3e4bf3dc34b6cf8b" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.692956 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-kkfbd"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.723726 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-kkfbd"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.736151 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.752814 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-d90b-account-create-lgjcg"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.756914 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.768706 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-d90b-account-create-lgjcg"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.779794 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.780171 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" containerID="cri-o://ef53fbe1eb381437837cbf0ca10ca7d3c3b0550ab75c3531ca2b8207894fbe27" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.780108 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" containerID="cri-o://65b14e1c027a8cae67975b1fce4a4009ca6aec0752d8342f09ee93d4ac090487" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.784280 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.798372 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.798917 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" containerID="cri-o://7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" gracePeriod=60 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.808399 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5b96cbd67f-h4t2t"] Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.809263 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data kube-api-access-sdcgs], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/heat-engine-5b96cbd67f-h4t2t" podUID="9330bbd8-2537-493f-9bb8-7de57ba8d350" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.836198 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.836431 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" containerID="cri-o://b5014c9bfac19f9acded35ea492bf1ac02255d3437039504d0647dddd1fdb140" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.837195 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" containerID="cri-o://d4b9ecf9551b7d71f24978b3596fb93c674a49dbabf2d694c982d16be6b892f8" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.856736 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret\") pod \"e216b65d-5bc3-453c-aa74-b9792e47575e\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.856903 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcjht\" (UniqueName: \"kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht\") pod \"e216b65d-5bc3-453c-aa74-b9792e47575e\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.856959 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle\") pod \"e216b65d-5bc3-453c-aa74-b9792e47575e\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.857029 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config\") pod \"e216b65d-5bc3-453c-aa74-b9792e47575e\" (UID: \"e216b65d-5bc3-453c-aa74-b9792e47575e\") " Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.889006 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht" (OuterVolumeSpecName: "kube-api-access-gcjht") pod "e216b65d-5bc3-453c-aa74-b9792e47575e" (UID: "e216b65d-5bc3-453c-aa74-b9792e47575e"). InnerVolumeSpecName "kube-api-access-gcjht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.898682 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.918804 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.919081 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5bb7c669fd-grgkd" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api-log" containerID="cri-o://c6b8156153448df4ea7ab21d1a8bb3213deefc4aed5b741b5a54c1beb1983b69" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.919505 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5bb7c669fd-grgkd" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api" containerID="cri-o://652fef03d0063382f0ac1925aeb8888b7617bf80e601c16452dfdee4b721b2c2" gracePeriod=30 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.934786 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.935364 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-576557876d-fngj4" podUID="eba7274f-a215-4784-a540-06013b44a29c" containerName="heat-api" containerID="cri-o://c337d9341511431c4a0f5f57944ece931c9e13a68e5f6796f87487029392bc65" gracePeriod=60 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.945780 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.946049 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" podUID="642159cf-6607-450f-84b9-f8ac137d173d" containerName="heat-cfnapi" containerID="cri-o://8e5bc391462c1807435f3141698540e3dbc744434e432fffc72f05920e38ffb9" gracePeriod=60 Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.967461 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcjht\" (UniqueName: \"kubernetes.io/projected/e216b65d-5bc3-453c-aa74-b9792e47575e-kube-api-access-gcjht\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[5018]: I1014 09:02:22.985227 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59d586bc68-bvmll"] Oct 14 09:02:22 crc kubenswrapper[5018]: E1014 09:02:22.986788 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data kube-api-access-lhqcc], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/heat-api-59d586bc68-bvmll" podUID="ab0106bb-211e-43ca-b0aa-0aa76099a4d2" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.065214 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-977pq"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.097578 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.165642 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-977pq"] Oct 14 09:02:23 crc kubenswrapper[5018]: W1014 09:02:23.179202 5018 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeab7a625_e4b8_4038_b9a7_ce1f7af786bf.slice/crio-b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d WatchSource:0}: Error finding container b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d: Status 404 returned error can't find the container with id b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.181074 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e216b65d-5bc3-453c-aa74-b9792e47575e" (UID: "e216b65d-5bc3-453c-aa74-b9792e47575e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.189813 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-033c-account-create-ptmqh"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.194625 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.197692 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e216b65d-5bc3-453c-aa74-b9792e47575e" (UID: "e216b65d-5bc3-453c-aa74-b9792e47575e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.209797 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-033c-account-create-ptmqh"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.217789 5018 generic.go:334] "Generic (PLEG): container finished" podID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerID="8ea98fcf3c27a68c727221403d77e0a9e5e402b4f567534f6bb06689161fc3eb" exitCode=0 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.218728 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerDied","Data":"8ea98fcf3c27a68c727221403d77e0a9e5e402b4f567534f6bb06689161fc3eb"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.223812 5018 generic.go:334] "Generic (PLEG): container finished" podID="32a1c03f-068c-4278-b239-480514f4b60d" containerID="0e6d9ef3c618844329be04c22cf07c96416cade5d7551e947ded7ff89d6b027d" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.223893 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerDied","Data":"0e6d9ef3c618844329be04c22cf07c96416cade5d7551e947ded7ff89d6b027d"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.225049 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.225558 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9830a080-9886-4018-b68c-60bda6dd6333" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb" gracePeriod=30 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.228944 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e216b65d-5bc3-453c-aa74-b9792e47575e" (UID: "e216b65d-5bc3-453c-aa74-b9792e47575e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.238901 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="galera" containerID="cri-o://005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" gracePeriod=30 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.239093 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.307866 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e216b65d-5bc3-453c-aa74-b9792e47575e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.307895 5018 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e216b65d-5bc3-453c-aa74-b9792e47575e-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.307992 5018 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.308043 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:25.308028549 +0000 UTC m=+7961.892075176 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-worker-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.308438 5018 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.308483 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:25.308475481 +0000 UTC m=+7961.892522108 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.308545 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.308743 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerName="nova-cell1-conductor-conductor" containerID="cri-o://9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" gracePeriod=30 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.322751 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8f5h4"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.336374 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8f5h4"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.345176 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.345362 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerName="nova-cell0-conductor-conductor" containerID="cri-o://43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" gracePeriod=30 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.370678 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.371199 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="prometheus" containerID="cri-o://5475b3ac1ffb8d1aa80f71e6a95e7eed4b52f74bd503013962bda9fef820ce2f" gracePeriod=600 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.371607 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="thanos-sidecar" containerID="cri-o://77311f9821ca7a858e16455376cf6227d93c7f441b52752f6083c5cada76a306" gracePeriod=600 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.371674 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="config-reloader" containerID="cri-o://cf8f965d2cb92a7c4505cb3c64420b7047da837b854170fa012a0781465f8cb5" gracePeriod=600 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.383099 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.383339 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/alertmanager-metric-storage-0" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="alertmanager" containerID="cri-o://ae3477a4359c14611282e48f047572b99c17263c4ab7fa2f6f48060c5cbbd512" gracePeriod=120 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.384488 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/alertmanager-metric-storage-0" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="config-reloader" containerID="cri-o://f8e67da18c07a0242bb09dc5c3e1ef768c85087115a502b92e1ccb4e2c0ad2a7" gracePeriod=120 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.399544 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.401730 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" containerID="cri-o://bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" gracePeriod=30 Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.420741 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15 is running failed: container process not found" containerID="fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.422545 5018 generic.go:334] "Generic (PLEG): container finished" podID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerID="65b14e1c027a8cae67975b1fce4a4009ca6aec0752d8342f09ee93d4ac090487" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.422643 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerDied","Data":"65b14e1c027a8cae67975b1fce4a4009ca6aec0752d8342f09ee93d4ac090487"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.447001 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.447483 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15 is running failed: container process not found" containerID="fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.450028 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15 is running failed: container process not found" containerID="fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.450104 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="ovsdbserver-sb" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.452341 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrond3e5-account-delete-4mrl8" event={"ID":"eab7a625-e4b8-4038-b9a7-ce1f7af786bf","Type":"ContainerStarted","Data":"b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.455938 5018 generic.go:334] "Generic (PLEG): container finished" podID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerID="b153baeb0241dacb32e954510d08cddc263a2d3ee9e254846604179d4e462655" exitCode=0 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.456033 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerDied","Data":"b153baeb0241dacb32e954510d08cddc263a2d3ee9e254846604179d4e462655"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.498002 5018 generic.go:334] "Generic (PLEG): container finished" podID="07df2031-64c4-4e6b-b70a-831edefc4468" containerID="ce612af3722b887703ef66d47d90f1bdbe0b44f623bc74647fe5ba7231238eed" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.498063 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerDied","Data":"ce612af3722b887703ef66d47d90f1bdbe0b44f623bc74647fe5ba7231238eed"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.538648 5018 generic.go:334] "Generic (PLEG): container finished" podID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerID="c6b8156153448df4ea7ab21d1a8bb3213deefc4aed5b741b5a54c1beb1983b69" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.538719 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerDied","Data":"c6b8156153448df4ea7ab21d1a8bb3213deefc4aed5b741b5a54c1beb1983b69"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.566192 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.566598 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655df9d575-9rwfg" event={"ID":"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e","Type":"ContainerDied","Data":"0f06a849dc9a8e2dca36450a84975071fc78612e587deb6b7312e11a7085a9f2"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.566664 5018 scope.go:117] "RemoveContainer" containerID="5045a096ebb1e46ec923b34da32996ecc7663e4e86bfba2ce6bb474b4f341eec" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.583168 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.598787 5018 generic.go:334] "Generic (PLEG): container finished" podID="827969fb-cda2-4793-8e73-51b2d1159caa" containerID="b10bff320cd1aef3f540117705e3e341d22490e36c13290cfba7eec50e8f6a71" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.598884 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerDied","Data":"b10bff320cd1aef3f540117705e3e341d22490e36c13290cfba7eec50e8f6a71"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.620293 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb\") pod \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.620518 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rwbw\" (UniqueName: \"kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw\") pod \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.620543 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc\") pod \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.620637 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config\") pod \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.620665 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb\") pod \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\" (UID: \"f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.623261 5018 generic.go:334] "Generic (PLEG): container finished" podID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerID="e8b14272735d2836638396ff83b7ead777ae1c594b519c1e3e4bf3dc34b6cf8b" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.623345 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerDied","Data":"e8b14272735d2836638396ff83b7ead777ae1c594b519c1e3e4bf3dc34b6cf8b"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.645958 5018 generic.go:334] "Generic (PLEG): container finished" podID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerID="9b40efcd93a346933415044024e90ede001495fcfd8fb0320885bab4c64fda7d" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.646251 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerDied","Data":"9b40efcd93a346933415044024e90ede001495fcfd8fb0320885bab4c64fda7d"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.668864 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw" (OuterVolumeSpecName: "kube-api-access-8rwbw") pod "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" (UID: "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e"). InnerVolumeSpecName "kube-api-access-8rwbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.669139 5018 scope.go:117] "RemoveContainer" containerID="1062635fc0651ed9abaa34283677781a36047d6343e5d6745ad4efbf76603dce" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.678700 5018 generic.go:334] "Generic (PLEG): container finished" podID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerID="b5014c9bfac19f9acded35ea492bf1ac02255d3437039504d0647dddd1fdb140" exitCode=143 Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.678793 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerDied","Data":"b5014c9bfac19f9acded35ea492bf1ac02255d3437039504d0647dddd1fdb140"} Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.678813 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.678875 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.709900 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.711825 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9703a145-2c14-4a43-a767-e6def8fe6063/ovsdbserver-sb/0.log" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.711893 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.714827 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.723901 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rwbw\" (UniqueName: \"kubernetes.io/projected/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-kube-api-access-8rwbw\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.724711 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.724745 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:27.724733402 +0000 UTC m=+7964.308780029 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data") pod "rabbitmq-cell1-server-0" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.752708 5018 scope.go:117] "RemoveContainer" containerID="ae4fa989c0d49cc43c46cfbb1bf94f22382f8020c2ea5b17401082ece4256593" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.768492 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" (UID: "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.789400 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" (UID: "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.807386 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b67bc93c-1c02-47fa-b2a5-57560256e89b/ovsdbserver-nb/0.log" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.807459 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.820255 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.839200 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle\") pod \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.839241 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.839263 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbrg7\" (UniqueName: \"kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.839304 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.843755 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846710 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846786 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846827 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846882 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846903 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom\") pod \"9330bbd8-2537-493f-9bb8-7de57ba8d350\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846929 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846952 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle\") pod \"9330bbd8-2537-493f-9bb8-7de57ba8d350\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.846987 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz9tr\" (UniqueName: \"kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847025 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847092 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs\") pod \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847121 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847139 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847158 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom\") pod \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847178 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847207 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs\") pod \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847225 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847246 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle\") pod \"9703a145-2c14-4a43-a767-e6def8fe6063\" (UID: \"9703a145-2c14-4a43-a767-e6def8fe6063\") " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.847991 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.848004 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.854884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts" (OuterVolumeSpecName: "scripts") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.859335 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config" (OuterVolumeSpecName: "config") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.862534 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr" (OuterVolumeSpecName: "kube-api-access-gz9tr") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "kube-api-access-gz9tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.862221 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ab0106bb-211e-43ca-b0aa-0aa76099a4d2" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.864744 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.865107 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.865547 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts" (OuterVolumeSpecName: "scripts") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.867866 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config" (OuterVolumeSpecName: "config") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.870175 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9330bbd8-2537-493f-9bb8-7de57ba8d350" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.870576 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ab0106bb-211e-43ca-b0aa-0aa76099a4d2" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.870790 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ab0106bb-211e-43ca-b0aa-0aa76099a4d2" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.873844 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7" (OuterVolumeSpecName: "kube-api-access-gbrg7") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "kube-api-access-gbrg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.882064 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.890089 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9330bbd8-2537-493f-9bb8-7de57ba8d350" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.895884 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab0106bb-211e-43ca-b0aa-0aa76099a4d2" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.899184 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" (UID: "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.906287 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "pvc-53f62397-4285-4aa0-a957-3300af156cb7". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.950958 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951019 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") pod \"heat-engine-5b96cbd67f-h4t2t\" (UID: \"9330bbd8-2537-493f-9bb8-7de57ba8d350\") " pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951231 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951245 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951256 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951268 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951278 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951289 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbrg7\" (UniqueName: \"kubernetes.io/projected/9703a145-2c14-4a43-a767-e6def8fe6063-kube-api-access-gbrg7\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951301 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951326 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") on node \"crc\" " Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951337 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951351 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67bc93c-1c02-47fa-b2a5-57560256e89b-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951362 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951373 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951383 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz9tr\" (UniqueName: \"kubernetes.io/projected/b67bc93c-1c02-47fa-b2a5-57560256e89b-kube-api-access-gz9tr\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951395 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9703a145-2c14-4a43-a767-e6def8fe6063-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951406 5018 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.951417 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.952317 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.952380 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:27.952362702 +0000 UTC m=+7964.536409329 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : secret "heat-config-data" not found Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.953445 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config" (OuterVolumeSpecName: "config") pod "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" (UID: "f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.958694 5018 projected.go:194] Error preparing data for projected volume kube-api-access-sdcgs for pod openstack/heat-engine-5b96cbd67f-h4t2t: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.958748 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs podName:9330bbd8-2537-493f-9bb8-7de57ba8d350 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:27.958733263 +0000 UTC m=+7964.542779890 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-sdcgs" (UniqueName: "kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs") pod "heat-engine-5b96cbd67f-h4t2t" (UID: "9330bbd8-2537-493f-9bb8-7de57ba8d350") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.967806 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: I1014 09:02:23.975855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:23 crc kubenswrapper[5018]: E1014 09:02:23.989585 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a podName:b67bc93c-1c02-47fa-b2a5-57560256e89b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:24.489561029 +0000 UTC m=+7961.073607656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "ovndbcluster-nb-etc-ovn" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.057940 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="9830a080-9886-4018-b68c-60bda6dd6333" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.1.104:6080/vnc_lite.html\": dial tcp 10.217.1.104:6080: connect: connection refused" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.058854 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.059125 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.059143 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.059153 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.059229 5018 secret.go:188] Couldn't get secret openstack/heat-config-data: secret "heat-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.059273 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.059258141 +0000 UTC m=+7964.643304758 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : secret "heat-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.073374 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.073564 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-53f62397-4285-4aa0-a957-3300af156cb7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7") on node "crc" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.085641 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.163091 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") pod \"heat-api-59d586bc68-bvmll\" (UID: \"ab0106bb-211e-43ca-b0aa-0aa76099a4d2\") " pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.164115 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.164153 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-53f62397-4285-4aa0-a957-3300af156cb7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-53f62397-4285-4aa0-a957-3300af156cb7\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.166267 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.168822 5018 projected.go:194] Error preparing data for projected volume kube-api-access-lhqcc for pod openstack/heat-api-59d586bc68-bvmll: failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.168894 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc podName:ab0106bb-211e-43ca-b0aa-0aa76099a4d2 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.168874886 +0000 UTC m=+7964.752921503 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-lhqcc" (UniqueName: "kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc") pod "heat-api-59d586bc68-bvmll" (UID: "ab0106bb-211e-43ca-b0aa-0aa76099a4d2") : failed to fetch token: serviceaccounts "heat-heat" not found Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.183828 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.245753 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "9703a145-2c14-4a43-a767-e6def8fe6063" (UID: "9703a145-2c14-4a43-a767-e6def8fe6063"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.268557 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.268685 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703a145-2c14-4a43-a767-e6def8fe6063-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.268697 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b67bc93c-1c02-47fa-b2a5-57560256e89b-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.268713 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.268780 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data podName:29481fb2-99c9-41f0-b650-a971fa5ff28b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.268760125 +0000 UTC m=+7964.852806752 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data") pod "rabbitmq-server-0" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b") : configmap "rabbitmq-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.320268 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.330377 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378267 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378296 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378352 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.37833485 +0000 UTC m=+7964.962381477 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378382 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.378364781 +0000 UTC m=+7964.962411408 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378836 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.378880 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:28.378871595 +0000 UTC m=+7964.962918222 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.582935 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") pod \"b67bc93c-1c02-47fa-b2a5-57560256e89b\" (UID: \"b67bc93c-1c02-47fa-b2a5-57560256e89b\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.602019 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "b67bc93c-1c02-47fa-b2a5-57560256e89b" (UID: "b67bc93c-1c02-47fa-b2a5-57560256e89b"). InnerVolumeSpecName "pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.602183 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400 is running failed: container process not found" containerID="9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.602537 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400 is running failed: container process not found" containerID="9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.602964 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400 is running failed: container process not found" containerID="9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.603019 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerName="nova-cell1-conductor-conductor" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.606584 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.606881 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.628970 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be" path="/var/lib/kubelet/pods/2fa8c3b8-96f8-4b5d-a1b9-ba443c09f5be/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.630382 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73aaf538-652f-43c9-90f2-ad5a844733dc" path="/var/lib/kubelet/pods/73aaf538-652f-43c9-90f2-ad5a844733dc/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.633083 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e54d5b-185d-40e9-a3a7-7c848eac44b1" path="/var/lib/kubelet/pods/90e54d5b-185d-40e9-a3a7-7c848eac44b1/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.633979 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0482650-95be-459a-901b-ee5ff6cb9e09" path="/var/lib/kubelet/pods/d0482650-95be-459a-901b-ee5ff6cb9e09/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.635252 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6927f2-a17e-4f85-be65-ea18d06166f5" path="/var/lib/kubelet/pods/db6927f2-a17e-4f85-be65-ea18d06166f5/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.635859 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e216b65d-5bc3-453c-aa74-b9792e47575e" path="/var/lib/kubelet/pods/e216b65d-5bc3-453c-aa74-b9792e47575e/volumes" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.685209 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") on node \"crc\" " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.702256 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.719229 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican3632-account-delete-sgfbk" event={"ID":"8fb6f85b-5866-46bf-a826-0a22243318c1","Type":"ContainerStarted","Data":"15a91fc6e98a111d8f9278bb335868fb696b025ffed40bedb18408ef675f0888"} Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.724194 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711 is running failed: container process not found" containerID="005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.724340 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heatd90b-account-delete-sh6hm" event={"ID":"90c10727-d872-4046-876c-39475c9d8623","Type":"ContainerStarted","Data":"cfe6a6bea837f6350f884c109dd8566794111b3168f22cf5979fecca71b15957"} Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.725068 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711 is running failed: container process not found" containerID="005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.725524 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711 is running failed: container process not found" containerID="005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.725555 5018 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711 is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="galera" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.726578 5018 generic.go:334] "Generic (PLEG): container finished" podID="eab7a625-e4b8-4038-b9a7-ce1f7af786bf" containerID="481f19ebdb8ac6219312b39cc0c7c557c520b379bcde4d8b5b13ed15fc368282" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.726663 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrond3e5-account-delete-4mrl8" event={"ID":"eab7a625-e4b8-4038-b9a7-ce1f7af786bf","Type":"ContainerDied","Data":"481f19ebdb8ac6219312b39cc0c7c557c520b379bcde4d8b5b13ed15fc368282"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.732608 5018 generic.go:334] "Generic (PLEG): container finished" podID="9830a080-9886-4018-b68c-60bda6dd6333" containerID="182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.732674 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9830a080-9886-4018-b68c-60bda6dd6333","Type":"ContainerDied","Data":"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.732703 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9830a080-9886-4018-b68c-60bda6dd6333","Type":"ContainerDied","Data":"df3d3f9450e3982d493280cbc646c2305b11a691c73e595e30de2b0b3307f2d4"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.732720 5018 scope.go:117] "RemoveContainer" containerID="182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.732809 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.740285 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.740425 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a") on node "crc" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.743639 5018 generic.go:334] "Generic (PLEG): container finished" podID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerID="f8e67da18c07a0242bb09dc5c3e1ef768c85087115a502b92e1ccb4e2c0ad2a7" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.743672 5018 generic.go:334] "Generic (PLEG): container finished" podID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerID="ae3477a4359c14611282e48f047572b99c17263c4ab7fa2f6f48060c5cbbd512" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.743699 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerDied","Data":"f8e67da18c07a0242bb09dc5c3e1ef768c85087115a502b92e1ccb4e2c0ad2a7"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.743759 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerDied","Data":"ae3477a4359c14611282e48f047572b99c17263c4ab7fa2f6f48060c5cbbd512"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.745949 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement85f4-account-delete-lx2cs" event={"ID":"5876dcf2-c564-45ac-bf38-bebf49e5661f","Type":"ContainerStarted","Data":"98061215156f5ec989c15434fba1483e8d6c14becbf30b9346f581c4159322af"} Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.747490 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.747950 5018 generic.go:334] "Generic (PLEG): container finished" podID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerID="9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.747997 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46a6263a-f328-43d0-8fcd-fb3610e88c30","Type":"ContainerDied","Data":"9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400"} Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.748685 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.749722 5018 generic.go:334] "Generic (PLEG): container finished" podID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerID="f3bfacf0b50c0a9351a97df8cd9a4d4d20587d8f3b2250add56a2512071e49f6" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.749763 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerDied","Data":"f3bfacf0b50c0a9351a97df8cd9a4d4d20587d8f3b2250add56a2512071e49f6"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.749778 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-559f8b448b-zgvh5" event={"ID":"521da8cb-e7fb-42b0-aa61-ee86ec329fad","Type":"ContainerDied","Data":"cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.749788 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc800375b687b525cd523caca0d7056e64cbf1fbca1a3daecd7319062536fa9" Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.750122 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:24 crc kubenswrapper[5018]: E1014 09:02:24.750149 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.754512 5018 generic.go:334] "Generic (PLEG): container finished" podID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerID="ac50df93bbc18e55844ca557b4aa55078662999d23f67f0d58048a5477b2f20c" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.754570 5018 generic.go:334] "Generic (PLEG): container finished" podID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerID="3c0372fbfcb7d77786b28550cb757bcf2d36775c128fc8d61c2eeed2bb4a27a4" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.754591 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerDied","Data":"ac50df93bbc18e55844ca557b4aa55078662999d23f67f0d58048a5477b2f20c"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.754643 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerDied","Data":"3c0372fbfcb7d77786b28550cb757bcf2d36775c128fc8d61c2eeed2bb4a27a4"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.756731 5018 generic.go:334] "Generic (PLEG): container finished" podID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerID="a7585a2f1c37963c5dbf63572362ad5333d952de5f762503637a079285e5bd70" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.756751 5018 generic.go:334] "Generic (PLEG): container finished" podID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerID="2d821f28ca619425ae5555a780634062a928277adbe2e77a11f5825e3ed63b9d" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.756781 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerDied","Data":"a7585a2f1c37963c5dbf63572362ad5333d952de5f762503637a079285e5bd70"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.756795 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerDied","Data":"2d821f28ca619425ae5555a780634062a928277adbe2e77a11f5825e3ed63b9d"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.757968 5018 generic.go:334] "Generic (PLEG): container finished" podID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerID="005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.758005 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerDied","Data":"005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.765678 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9703a145-2c14-4a43-a767-e6def8fe6063/ovsdbserver-sb/0.log" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.765778 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.760127 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance72c6-account-delete-27sjg" event={"ID":"7f55eb6e-96bb-4019-83a9-10ba9f455ecc","Type":"ContainerStarted","Data":"3c33d9583bae5f2c62f8c334a794707b142bb53b5e990bfcdba248d32775bc34"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.768990 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9703a145-2c14-4a43-a767-e6def8fe6063","Type":"ContainerDied","Data":"8f33b654947a44e55ec8833c78e946654d9bb0accc7e767d8e0ef9ab5ac39e41"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.783852 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b67bc93c-1c02-47fa-b2a5-57560256e89b/ovsdbserver-nb/0.log" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.783928 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b67bc93c-1c02-47fa-b2a5-57560256e89b","Type":"ContainerDied","Data":"0bd72932314fd0730fac8cd026bc5e129bd1070837a76c295efa99938592dd2d"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.784004 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.793421 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44ff4af5-9f43-4e97-af81-206fcf9e123a\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.795736 5018 generic.go:334] "Generic (PLEG): container finished" podID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerID="d43f95354a677e6c3aafe9f165dbe59e283b3bddc5f2df6c04584560150ca6b8" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.795795 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerDied","Data":"d43f95354a677e6c3aafe9f165dbe59e283b3bddc5f2df6c04584560150ca6b8"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.798577 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerID="77311f9821ca7a858e16455376cf6227d93c7f441b52752f6083c5cada76a306" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.798599 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerID="cf8f965d2cb92a7c4505cb3c64420b7047da837b854170fa012a0781465f8cb5" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.798607 5018 generic.go:334] "Generic (PLEG): container finished" podID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerID="5475b3ac1ffb8d1aa80f71e6a95e7eed4b52f74bd503013962bda9fef820ce2f" exitCode=0 Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.798670 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59d586bc68-bvmll" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804094 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerDied","Data":"77311f9821ca7a858e16455376cf6227d93c7f441b52752f6083c5cada76a306"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804148 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5b96cbd67f-h4t2t" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804171 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerDied","Data":"cf8f965d2cb92a7c4505cb3c64420b7047da837b854170fa012a0781465f8cb5"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804187 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerDied","Data":"5475b3ac1ffb8d1aa80f71e6a95e7eed4b52f74bd503013962bda9fef820ce2f"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804199 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab","Type":"ContainerDied","Data":"917dcfbfd44e3b51a2c06de6a38b735b9a862ee3bfad2604b0b2b3961b0e967b"} Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.804214 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="917dcfbfd44e3b51a2c06de6a38b735b9a862ee3bfad2604b0b2b3961b0e967b" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.869134 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.900575 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6g2g\" (UniqueName: \"kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g\") pod \"9830a080-9886-4018-b68c-60bda6dd6333\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.900786 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs\") pod \"9830a080-9886-4018-b68c-60bda6dd6333\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.900868 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle\") pod \"9830a080-9886-4018-b68c-60bda6dd6333\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.900930 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs\") pod \"9830a080-9886-4018-b68c-60bda6dd6333\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.900952 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data\") pod \"9830a080-9886-4018-b68c-60bda6dd6333\" (UID: \"9830a080-9886-4018-b68c-60bda6dd6333\") " Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.948226 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g" (OuterVolumeSpecName: "kube-api-access-x6g2g") pod "9830a080-9886-4018-b68c-60bda6dd6333" (UID: "9830a080-9886-4018-b68c-60bda6dd6333"). InnerVolumeSpecName "kube-api-access-x6g2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.954334 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data" (OuterVolumeSpecName: "config-data") pod "9830a080-9886-4018-b68c-60bda6dd6333" (UID: "9830a080-9886-4018-b68c-60bda6dd6333"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.954403 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9830a080-9886-4018-b68c-60bda6dd6333" (UID: "9830a080-9886-4018-b68c-60bda6dd6333"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:24 crc kubenswrapper[5018]: I1014 09:02:24.999263 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "9830a080-9886-4018-b68c-60bda6dd6333" (UID: "9830a080-9886-4018-b68c-60bda6dd6333"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004499 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xjbm\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004689 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004727 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004777 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004794 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004865 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004889 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.004912 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs\") pod \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\" (UID: \"521da8cb-e7fb-42b0-aa61-ee86ec329fad\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.005734 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6g2g\" (UniqueName: \"kubernetes.io/projected/9830a080-9886-4018-b68c-60bda6dd6333-kube-api-access-x6g2g\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.005753 5018 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.005761 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.005772 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.007113 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.007249 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.014749 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "9830a080-9886-4018-b68c-60bda6dd6333" (UID: "9830a080-9886-4018-b68c-60bda6dd6333"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.016944 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm" (OuterVolumeSpecName: "kube-api-access-5xjbm") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "kube-api-access-5xjbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.023166 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.035774 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.037176 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-central-agent" containerID="cri-o://68ece5d157ffc84bb5586ef4fa04f0d496bf0dbee2e3cc8cbcd837a81d450889" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.037559 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="proxy-httpd" containerID="cri-o://82fa2b51a89e84f109bc39c3a296ac0d8bc72f1de856de23122ba2cda3338fdd" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.037601 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="sg-core" containerID="cri-o://087e0937d73febf55e6d268d5a2a924ae960ed5b8f8408710ab27cc2646a8605" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.037657 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-notification-agent" containerID="cri-o://298b04d7b493d03a2ba6b0a665cc5d3190695959aea3d6fec443b33a4c7ab1c7" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.063864 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.064054 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerName="kube-state-metrics" containerID="cri-o://0835e3a86cc1bed0b46ea89d60d79c970f2f5d7aa91a959459086c75a2a30e00" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.108238 5018 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.108269 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.108278 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/521da8cb-e7fb-42b0-aa61-ee86ec329fad-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.108287 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xjbm\" (UniqueName: \"kubernetes.io/projected/521da8cb-e7fb-42b0-aa61-ee86ec329fad-kube-api-access-5xjbm\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.108298 5018 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9830a080-9886-4018-b68c-60bda6dd6333-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.157960 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.158363 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" containerName="memcached" containerID="cri-o://49e3187ba65a221c522e49e4c2118872d506a95aa0ad3018a57afeb0f6482e51" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.204402 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.215345 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.255422 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.281937 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.282158 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5479495b9c-w7q9b" podUID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" containerName="keystone-api" containerID="cri-o://3b364f7b933d0c59e076d99185b783f5158d326517a316e15f35f3ba0599def5" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.289700 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cron-29340541-h7n4r"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.294634 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cron-29340541-h7n4r"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.297065 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.302843 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data" (OuterVolumeSpecName: "config-data") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.313919 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "521da8cb-e7fb-42b0-aa61-ee86ec329fad" (UID: "521da8cb-e7fb-42b0-aa61-ee86ec329fad"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: E1014 09:02:25.348067 5018 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.348348 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: E1014 09:02:25.348372 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:29.348356172 +0000 UTC m=+7965.932402799 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-worker-config-data" not found Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.348388 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.348400 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/521da8cb-e7fb-42b0-aa61-ee86ec329fad-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: E1014 09:02:25.348403 5018 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 14 09:02:25 crc kubenswrapper[5018]: E1014 09:02:25.348429 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:29.348423584 +0000 UTC m=+7965.932470211 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-config-data" not found Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.394657 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.395920 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.399776 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-655df9d575-9rwfg"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.431562 5018 scope.go:117] "RemoveContainer" containerID="182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb" Oct 14 09:02:25 crc kubenswrapper[5018]: E1014 09:02:25.439807 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb\": container with ID starting with 182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb not found: ID does not exist" containerID="182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.439865 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb"} err="failed to get container status \"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb\": rpc error: code = NotFound desc = could not find container \"182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb\": container with ID starting with 182fab42b5d27e67356932f07905755c8edeb95eca4b735563e3928fd0e4e4eb not found: ID does not exist" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.439890 5018 scope.go:117] "RemoveContainer" containerID="b8af3ead5cf9c4d299de3975f6ab77037e7e1da24c12c876f67ca25674008762" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.449241 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.449445 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ctxl\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.449589 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.449766 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.449929 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450036 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450128 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450316 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450423 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450552 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.450713 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out\") pod \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\" (UID: \"ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.456052 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.457500 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config" (OuterVolumeSpecName: "config") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.458222 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl" (OuterVolumeSpecName: "kube-api-access-2ctxl") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "kube-api-access-2ctxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.460110 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out" (OuterVolumeSpecName: "config-out") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.460220 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.466698 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.469355 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.473868 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.474283 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.494512 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.520106 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.534081 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.537005 5018 scope.go:117] "RemoveContainer" containerID="fc3af86cc64613033ee49c65bc9d48332477bf4b3c09bf341b95ec38b649bb15" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.549819 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.552471 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.553278 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data\") pod \"46a6263a-f328-43d0-8fcd-fb3610e88c30\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.553428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle\") pod \"46a6263a-f328-43d0-8fcd-fb3610e88c30\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.553508 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v64vt\" (UniqueName: \"kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt\") pod \"46a6263a-f328-43d0-8fcd-fb3610e88c30\" (UID: \"46a6263a-f328-43d0-8fcd-fb3610e88c30\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.561201 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.562427 5018 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.563952 5018 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.564035 5018 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config-out\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.564105 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.564167 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ctxl\" (UniqueName: \"kubernetes.io/projected/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-kube-api-access-2ctxl\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.564935 5018 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.565029 5018 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.565094 5018 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.565172 5018 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.569599 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.573221 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt" (OuterVolumeSpecName: "kube-api-access-v64vt") pod "46a6263a-f328-43d0-8fcd-fb3610e88c30" (UID: "46a6263a-f328-43d0-8fcd-fb3610e88c30"). InnerVolumeSpecName "kube-api-access-v64vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.587585 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59d586bc68-bvmll"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.595819 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-59d586bc68-bvmll"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.615858 5018 scope.go:117] "RemoveContainer" containerID="e06750d462802f2f98ea3875ae3f637ed5b88aae6f70ae700e5336271d04d93c" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.616258 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.640903 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="galera" containerID="cri-o://589d1bdd70b59fd98fe13f906d2c0e56fa29052d702a35f68536bb9f4b51c629" gracePeriod=30 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.664107 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.665882 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh82b\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.665934 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666002 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666069 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666090 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqqt2\" (UniqueName: \"kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666127 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666154 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666209 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666244 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id\") pod \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\" (UID: \"6b5c5d38-56a1-447d-96f4-a8427ff3df82\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666288 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666320 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.666347 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db\") pod \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\" (UID: \"5f24304f-0a66-4c0f-b8cb-3341f9fb926d\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.667153 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.667172 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhqcc\" (UniqueName: \"kubernetes.io/projected/ab0106bb-211e-43ca-b0aa-0aa76099a4d2-kube-api-access-lhqcc\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.667191 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") on node \"crc\" " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.667202 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v64vt\" (UniqueName: \"kubernetes.io/projected/46a6263a-f328-43d0-8fcd-fb3610e88c30-kube-api-access-v64vt\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.668937 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5b96cbd67f-h4t2t"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.669304 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db" (OuterVolumeSpecName: "alertmanager-metric-storage-db") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "alertmanager-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.671481 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.688682 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.697429 5018 scope.go:117] "RemoveContainer" containerID="f7ec6a6c69b6da40d586a0fd10933c3e47ceb135c7cc71288eb38bcff403953a" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.697710 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.701851 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts" (OuterVolumeSpecName: "scripts") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.703068 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2" (OuterVolumeSpecName: "kube-api-access-hqqt2") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "kube-api-access-hqqt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.708193 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5b96cbd67f-h4t2t"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.711182 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out" (OuterVolumeSpecName: "config-out") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.727818 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume" (OuterVolumeSpecName: "config-volume") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.728560 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b" (OuterVolumeSpecName: "kube-api-access-jh82b") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "kube-api-access-jh82b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.742726 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.761775 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768508 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768662 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768739 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768768 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nljm8\" (UniqueName: \"kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768815 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768891 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.768969 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769598 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769653 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769675 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769670 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs" (OuterVolumeSpecName: "logs") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769769 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2kg9\" (UniqueName: \"kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.769795 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.770209 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.771305 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.771362 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.771427 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run\") pod \"fc347af0-75b9-4175-bc5e-70f3078e35c6\" (UID: \"fc347af0-75b9-4175-bc5e-70f3078e35c6\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.771466 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default\") pod \"8a5a6c5b-9410-4fde-98c4-112a882b366c\" (UID: \"8a5a6c5b-9410-4fde-98c4-112a882b366c\") " Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.771826 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.772320 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773061 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773091 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773110 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdcgs\" (UniqueName: \"kubernetes.io/projected/9330bbd8-2537-493f-9bb8-7de57ba8d350-kube-api-access-sdcgs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773125 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773139 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9330bbd8-2537-493f-9bb8-7de57ba8d350-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773151 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b5c5d38-56a1-447d-96f4-a8427ff3df82-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773164 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773176 5018 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773192 5018 reconciler_common.go:293] "Volume detached for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-alertmanager-metric-storage-db\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773205 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh82b\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-kube-api-access-jh82b\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773217 5018 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-config-out\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773229 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc347af0-75b9-4175-bc5e-70f3078e35c6-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773242 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqqt2\" (UniqueName: \"kubernetes.io/projected/6b5c5d38-56a1-447d-96f4-a8427ff3df82-kube-api-access-hqqt2\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.773253 5018 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.776168 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.779661 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.785699 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.807912 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets" (OuterVolumeSpecName: "secrets") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.808537 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9" (OuterVolumeSpecName: "kube-api-access-b2kg9") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "kube-api-access-b2kg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.808580 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8" (OuterVolumeSpecName: "kube-api-access-nljm8") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "kube-api-access-nljm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.808504 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts" (OuterVolumeSpecName: "scripts") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.822272 5018 generic.go:334] "Generic (PLEG): container finished" podID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerID="0835e3a86cc1bed0b46ea89d60d79c970f2f5d7aa91a959459086c75a2a30e00" exitCode=2 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.822355 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d48b611-d8d9-48a6-9179-d58ffa16ac18","Type":"ContainerDied","Data":"0835e3a86cc1bed0b46ea89d60d79c970f2f5d7aa91a959459086c75a2a30e00"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.826182 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.826203 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8a5a6c5b-9410-4fde-98c4-112a882b366c","Type":"ContainerDied","Data":"557b6127928f5f5395cc76a2bae8881ba30a62f5b03767fefc232ab4759c7403"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.826277 5018 scope.go:117] "RemoveContainer" containerID="005bfa12e1570e839a76fa4b5a63734f9597848834e1acf8d42834ca23a6c711" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.830338 5018 generic.go:334] "Generic (PLEG): container finished" podID="8fb6f85b-5866-46bf-a826-0a22243318c1" containerID="8d24ed0cc4e69b98105c8dc7d448104cba6f7918634697e95983418dae0608c6" exitCode=1 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.830506 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican3632-account-delete-sgfbk" event={"ID":"8fb6f85b-5866-46bf-a826-0a22243318c1","Type":"ContainerDied","Data":"8d24ed0cc4e69b98105c8dc7d448104cba6f7918634697e95983418dae0608c6"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858451 5018 generic.go:334] "Generic (PLEG): container finished" podID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerID="82fa2b51a89e84f109bc39c3a296ac0d8bc72f1de856de23122ba2cda3338fdd" exitCode=0 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858482 5018 generic.go:334] "Generic (PLEG): container finished" podID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerID="087e0937d73febf55e6d268d5a2a924ae960ed5b8f8408710ab27cc2646a8605" exitCode=2 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858491 5018 generic.go:334] "Generic (PLEG): container finished" podID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerID="68ece5d157ffc84bb5586ef4fa04f0d496bf0dbee2e3cc8cbcd837a81d450889" exitCode=0 Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858559 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerDied","Data":"82fa2b51a89e84f109bc39c3a296ac0d8bc72f1de856de23122ba2cda3338fdd"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858643 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerDied","Data":"087e0937d73febf55e6d268d5a2a924ae960ed5b8f8408710ab27cc2646a8605"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.858656 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerDied","Data":"68ece5d157ffc84bb5586ef4fa04f0d496bf0dbee2e3cc8cbcd837a81d450889"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.864294 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"5f24304f-0a66-4c0f-b8cb-3341f9fb926d","Type":"ContainerDied","Data":"4dfa3df785869f037ff7b347a4d536e19df442cb05356e13eeef22705f84f0bd"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.864397 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.867224 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46a6263a-f328-43d0-8fcd-fb3610e88c30","Type":"ContainerDied","Data":"e7ef6f4182de9825c89df33a03ee6d4c007d647ba422ba1d838a84bd08c311fe"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.867277 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877184 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877205 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877216 5018 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877230 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nljm8\" (UniqueName: \"kubernetes.io/projected/8a5a6c5b-9410-4fde-98c4-112a882b366c-kube-api-access-nljm8\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877241 5018 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5a6c5b-9410-4fde-98c4-112a882b366c-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877255 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2kg9\" (UniqueName: \"kubernetes.io/projected/fc347af0-75b9-4175-bc5e-70f3078e35c6-kube-api-access-b2kg9\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.877265 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a5a6c5b-9410-4fde-98c4-112a882b366c-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.881402 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b5c5d38-56a1-447d-96f4-a8427ff3df82","Type":"ContainerDied","Data":"bf7cca211ae8175bfa6033459b6e1fe68996cb35b76e89f90d6110438a73cc63"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.881543 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.883075 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.152:3000/\": dial tcp 10.217.1.152:3000: connect: connection refused" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.892478 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.892804 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc347af0-75b9-4175-bc5e-70f3078e35c6","Type":"ContainerDied","Data":"6aa0fb196e8892875265f3b590d03151715a0a279f51639d7b81ab50ccf134c4"} Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.892929 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:02:25 crc kubenswrapper[5018]: I1014 09:02:25.893012 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-559f8b448b-zgvh5" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:25.999744 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c" (OuterVolumeSpecName: "mysql-db") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "pvc-b099a311-342f-4a62-bd7e-22b2920fd33c". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.077756 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46a6263a-f328-43d0-8fcd-fb3610e88c30" (UID: "46a6263a-f328-43d0-8fcd-fb3610e88c30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.096344 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.096814 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") on node \"crc\" " Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.156511 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.206428 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.213876 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-576557876d-fngj4" podUID="eba7274f-a215-4784-a540-06013b44a29c" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.1.131:8004/healthcheck\": read tcp 10.217.0.2:36466->10.217.1.131:8004: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.221389 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.221575 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df") on node "crc" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.224244 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.270714 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.108:8775/\": read tcp 10.217.0.2:38442->10.217.1.108:8775: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.271218 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.108:8775/\": read tcp 10.217.0.2:38454->10.217.1.108:8775: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.278298 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data" (OuterVolumeSpecName: "config-data") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.299896 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.300130 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b099a311-342f-4a62-bd7e-22b2920fd33c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c") on node "crc" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.307432 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "8a5a6c5b-9410-4fde-98c4-112a882b366c" (UID: "8a5a6c5b-9410-4fde-98c4-112a882b366c"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.307980 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bf0118-09f1-4f5b-bc3e-b157a1bfb0df\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.308013 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.308023 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.308032 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b099a311-342f-4a62-bd7e-22b2920fd33c\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.308041 5018 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a5a6c5b-9410-4fde-98c4-112a882b366c-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.347813 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.349456 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.350745 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.350783 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerName="nova-cell0-conductor-conductor" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.351883 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fc347af0-75b9-4175-bc5e-70f3078e35c6" (UID: "fc347af0-75b9-4175-bc5e-70f3078e35c6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.384053 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.393332 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.396583 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 09:02:26 crc kubenswrapper[5018]: E1014 09:02:26.396766 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.408074 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data" (OuterVolumeSpecName: "config-data") pod "46a6263a-f328-43d0-8fcd-fb3610e88c30" (UID: "46a6263a-f328-43d0-8fcd-fb3610e88c30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.411039 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a6263a-f328-43d0-8fcd-fb3610e88c30-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.411207 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc347af0-75b9-4175-bc5e-70f3078e35c6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.442278 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5bb7c669fd-grgkd" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.1.50:9311/healthcheck\": read tcp 10.217.0.2:56316->10.217.1.50:9311: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.442325 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5bb7c669fd-grgkd" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.1.50:9311/healthcheck\": read tcp 10.217.0.2:56314->10.217.1.50:9311: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.445915 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config" (OuterVolumeSpecName: "web-config") pod "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" (UID: "ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.451694 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" podUID="642159cf-6607-450f-84b9-f8ac137d173d" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.130:8000/healthcheck\": read tcp 10.217.0.2:47862->10.217.1.130:8000: read: connection reset by peer" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.480718 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.509013 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.513015 5018 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab-web-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.513041 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.540551 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config" (OuterVolumeSpecName: "web-config") pod "5f24304f-0a66-4c0f-b8cb-3341f9fb926d" (UID: "5f24304f-0a66-4c0f-b8cb-3341f9fb926d"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.564606 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data" (OuterVolumeSpecName: "config-data") pod "6b5c5d38-56a1-447d-96f4-a8427ff3df82" (UID: "6b5c5d38-56a1-447d-96f4-a8427ff3df82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.614200 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5c5d38-56a1-447d-96f4-a8427ff3df82-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.614441 5018 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f24304f-0a66-4c0f-b8cb-3341f9fb926d-web-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.643556 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9330bbd8-2537-493f-9bb8-7de57ba8d350" path="/var/lib/kubelet/pods/9330bbd8-2537-493f-9bb8-7de57ba8d350/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.650173 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" path="/var/lib/kubelet/pods/9703a145-2c14-4a43-a767-e6def8fe6063/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.659169 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9830a080-9886-4018-b68c-60bda6dd6333" path="/var/lib/kubelet/pods/9830a080-9886-4018-b68c-60bda6dd6333/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.659967 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab0106bb-211e-43ca-b0aa-0aa76099a4d2" path="/var/lib/kubelet/pods/ab0106bb-211e-43ca-b0aa-0aa76099a4d2/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.660456 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" path="/var/lib/kubelet/pods/b67bc93c-1c02-47fa-b2a5-57560256e89b/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.671111 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f117dc78-c8fa-43ce-922c-6c035ff929ac" path="/var/lib/kubelet/pods/f117dc78-c8fa-43ce-922c-6c035ff929ac/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.678083 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" path="/var/lib/kubelet/pods/f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e/volumes" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.856470 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.112:8774/\": dial tcp 10.217.1.112:8774: connect: connection refused" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.857017 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.112:8774/\": dial tcp 10.217.1.112:8774: connect: connection refused" Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.957997 5018 generic.go:334] "Generic (PLEG): container finished" podID="2cd4a8e5-6715-4830-a0df-134660996996" containerID="392ee2ce4169eab5c3bfa508967090781008c6bb4b91e91bc63269bdd41562f3" exitCode=0 Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.958090 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerDied","Data":"392ee2ce4169eab5c3bfa508967090781008c6bb4b91e91bc63269bdd41562f3"} Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.972881 5018 generic.go:334] "Generic (PLEG): container finished" podID="90c10727-d872-4046-876c-39475c9d8623" containerID="ae59eb68f544fafc2469fbd82a6797652eaa3a4b90fe15cf3bd03e230ec1f737" exitCode=1 Oct 14 09:02:26 crc kubenswrapper[5018]: I1014 09:02:26.972944 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heatd90b-account-delete-sh6hm" event={"ID":"90c10727-d872-4046-876c-39475c9d8623","Type":"ContainerDied","Data":"ae59eb68f544fafc2469fbd82a6797652eaa3a4b90fe15cf3bd03e230ec1f737"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.011663 5018 generic.go:334] "Generic (PLEG): container finished" podID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerID="43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.011745 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fe75171-60d0-48df-a61a-c0e828f23f66","Type":"ContainerDied","Data":"43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.033353 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d48b611-d8d9-48a6-9179-d58ffa16ac18","Type":"ContainerDied","Data":"5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.033381 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aa13e688ccac3efd91bfb61388e3d47609d6cf39e1334838726c3df49f677c2" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.058273 5018 generic.go:334] "Generic (PLEG): container finished" podID="07df2031-64c4-4e6b-b70a-831edefc4468" containerID="cc378e4ea02d43f477ec8203ef137c674e7055a3b57a90d0bac877f9fdd800da" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.058353 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerDied","Data":"cc378e4ea02d43f477ec8203ef137c674e7055a3b57a90d0bac877f9fdd800da"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.058379 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07df2031-64c4-4e6b-b70a-831edefc4468","Type":"ContainerDied","Data":"4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.058389 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4650d763bb00871673ad279b4d147cd4ee7a656b2d02facc208f69e9d7620aec" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.064978 5018 generic.go:334] "Generic (PLEG): container finished" podID="642159cf-6607-450f-84b9-f8ac137d173d" containerID="8e5bc391462c1807435f3141698540e3dbc744434e432fffc72f05920e38ffb9" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.065040 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" event={"ID":"642159cf-6607-450f-84b9-f8ac137d173d","Type":"ContainerDied","Data":"8e5bc391462c1807435f3141698540e3dbc744434e432fffc72f05920e38ffb9"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.072055 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican3632-account-delete-sgfbk" event={"ID":"8fb6f85b-5866-46bf-a826-0a22243318c1","Type":"ContainerDied","Data":"15a91fc6e98a111d8f9278bb335868fb696b025ffed40bedb18408ef675f0888"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.072095 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15a91fc6e98a111d8f9278bb335868fb696b025ffed40bedb18408ef675f0888" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.075876 5018 generic.go:334] "Generic (PLEG): container finished" podID="827969fb-cda2-4793-8e73-51b2d1159caa" containerID="89c2c3c1c29bf6244a053cff8a3b895dd672b6f514f9287e56db1eeba83bd95e" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.075969 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerDied","Data":"89c2c3c1c29bf6244a053cff8a3b895dd672b6f514f9287e56db1eeba83bd95e"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.104186 5018 generic.go:334] "Generic (PLEG): container finished" podID="5876dcf2-c564-45ac-bf38-bebf49e5661f" containerID="677b825788e48fd7761329b03c95970c16bed163be1fd67aade0b8effcd2bbcc" exitCode=1 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.104256 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement85f4-account-delete-lx2cs" event={"ID":"5876dcf2-c564-45ac-bf38-bebf49e5661f","Type":"ContainerDied","Data":"677b825788e48fd7761329b03c95970c16bed163be1fd67aade0b8effcd2bbcc"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.138384 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrond3e5-account-delete-4mrl8" event={"ID":"eab7a625-e4b8-4038-b9a7-ce1f7af786bf","Type":"ContainerDied","Data":"b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.138418 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b64a64d80442435e30bb0e7fea3813f5fe312bc93c8a21398ddee95477c5422d" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.143094 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance72c6-account-delete-27sjg" event={"ID":"7f55eb6e-96bb-4019-83a9-10ba9f455ecc","Type":"ContainerDied","Data":"99764dec4ca9c95b5f6b94a2065141c0caf1d883b5b4424ac566e1a0775432f3"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.142975 5018 generic.go:334] "Generic (PLEG): container finished" podID="7f55eb6e-96bb-4019-83a9-10ba9f455ecc" containerID="99764dec4ca9c95b5f6b94a2065141c0caf1d883b5b4424ac566e1a0775432f3" exitCode=1 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.161022 5018 generic.go:334] "Generic (PLEG): container finished" podID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerID="ef53fbe1eb381437837cbf0ca10ca7d3c3b0550ab75c3531ca2b8207894fbe27" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.161103 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerDied","Data":"ef53fbe1eb381437837cbf0ca10ca7d3c3b0550ab75c3531ca2b8207894fbe27"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.164370 5018 generic.go:334] "Generic (PLEG): container finished" podID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerID="40fb8b68d8bb04c1d344475d4f00e45548e0b4c28983fcb23f3b012d39221907" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.164433 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerDied","Data":"40fb8b68d8bb04c1d344475d4f00e45548e0b4c28983fcb23f3b012d39221907"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.164455 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5b8df9c49b-fpwlm" event={"ID":"2a17b000-a726-4072-8b96-04f317ff7fb0","Type":"ContainerDied","Data":"1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.164467 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f1af294b7374c1dbb5e81e2742001ce6c7ca1dfa8c412d83578327c87a4782f" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.166332 5018 generic.go:334] "Generic (PLEG): container finished" podID="eba7274f-a215-4784-a540-06013b44a29c" containerID="c337d9341511431c4a0f5f57944ece931c9e13a68e5f6796f87487029392bc65" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.166390 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-576557876d-fngj4" event={"ID":"eba7274f-a215-4784-a540-06013b44a29c","Type":"ContainerDied","Data":"c337d9341511431c4a0f5f57944ece931c9e13a68e5f6796f87487029392bc65"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.169240 5018 generic.go:334] "Generic (PLEG): container finished" podID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerID="d4b9ecf9551b7d71f24978b3596fb93c674a49dbabf2d694c982d16be6b892f8" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.169291 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerDied","Data":"d4b9ecf9551b7d71f24978b3596fb93c674a49dbabf2d694c982d16be6b892f8"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.173018 5018 generic.go:334] "Generic (PLEG): container finished" podID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerID="652fef03d0063382f0ac1925aeb8888b7617bf80e601c16452dfdee4b721b2c2" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.173076 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerDied","Data":"652fef03d0063382f0ac1925aeb8888b7617bf80e601c16452dfdee4b721b2c2"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.189606 5018 generic.go:334] "Generic (PLEG): container finished" podID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerID="298b04d7b493d03a2ba6b0a665cc5d3190695959aea3d6fec443b33a4c7ab1c7" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.189689 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerDied","Data":"298b04d7b493d03a2ba6b0a665cc5d3190695959aea3d6fec443b33a4c7ab1c7"} Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.191984 5018 generic.go:334] "Generic (PLEG): container finished" podID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" containerID="49e3187ba65a221c522e49e4c2118872d506a95aa0ad3018a57afeb0f6482e51" exitCode=0 Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.192017 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"941d01f1-c45d-49c3-923b-d685ad3a2f4b","Type":"ContainerDied","Data":"49e3187ba65a221c522e49e4c2118872d506a95aa0ad3018a57afeb0f6482e51"} Oct 14 09:02:27 crc kubenswrapper[5018]: E1014 09:02:27.748666 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:27 crc kubenswrapper[5018]: E1014 09:02:27.749029 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:35.749014017 +0000 UTC m=+7972.333060644 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data") pod "rabbitmq-cell1-server-0" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : configmap "rabbitmq-cell1-config-data" not found Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.779698 5018 scope.go:117] "RemoveContainer" containerID="85f4670aa47a47b14a062924c76d871821892d2a11c13336e362bd97129914b3" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.838508 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.855040 5018 scope.go:117] "RemoveContainer" containerID="f8e67da18c07a0242bb09dc5c3e1ef768c85087115a502b92e1ccb4e2c0ad2a7" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.883851 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.901439 5018 scope.go:117] "RemoveContainer" containerID="ae3477a4359c14611282e48f047572b99c17263c4ab7fa2f6f48060c5cbbd512" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.906065 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.926587 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.938711 5018 scope.go:117] "RemoveContainer" containerID="93a6064e24381311dfd0a20d1a614392026b9ff3231be0c6e8ac754eeaa62f1c" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.944814 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.946776 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.952836 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.953532 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrcz4\" (UniqueName: \"kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4\") pod \"eab7a625-e4b8-4038-b9a7-ce1f7af786bf\" (UID: \"eab7a625-e4b8-4038-b9a7-ce1f7af786bf\") " Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.976532 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-576557876d-fngj4" Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.993303 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:27 crc kubenswrapper[5018]: I1014 09:02:27.995236 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.007608 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.009659 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4" (OuterVolumeSpecName: "kube-api-access-vrcz4") pod "eab7a625-e4b8-4038-b9a7-ce1f7af786bf" (UID: "eab7a625-e4b8-4038-b9a7-ce1f7af786bf"). InnerVolumeSpecName "kube-api-access-vrcz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.016667 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.023394 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.032333 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.054603 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.055475 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056339 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056368 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config\") pod \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056406 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056428 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056456 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056472 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056497 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gjhm\" (UniqueName: \"kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.056504 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057094 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs" (OuterVolumeSpecName: "logs") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057152 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057466 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057489 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057510 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle\") pod \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057530 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057580 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057629 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs\") pod \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057667 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057691 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057726 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6rwq\" (UniqueName: \"kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq\") pod \"8fb6f85b-5866-46bf-a826-0a22243318c1\" (UID: \"8fb6f85b-5866-46bf-a826-0a22243318c1\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057767 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrdp2\" (UniqueName: \"kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2\") pod \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\" (UID: \"3d48b611-d8d9-48a6-9179-d58ffa16ac18\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057789 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crpr6\" (UniqueName: \"kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6\") pod \"07df2031-64c4-4e6b-b70a-831edefc4468\" (UID: \"07df2031-64c4-4e6b-b70a-831edefc4468\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.057874 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts\") pod \"2a17b000-a726-4072-8b96-04f317ff7fb0\" (UID: \"2a17b000-a726-4072-8b96-04f317ff7fb0\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.058428 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrcz4\" (UniqueName: \"kubernetes.io/projected/eab7a625-e4b8-4038-b9a7-ce1f7af786bf-kube-api-access-vrcz4\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.058440 5018 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07df2031-64c4-4e6b-b70a-831edefc4468-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.058449 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a17b000-a726-4072-8b96-04f317ff7fb0-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.059875 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs" (OuterVolumeSpecName: "logs") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.062969 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts" (OuterVolumeSpecName: "scripts") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.064940 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.070717 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.075970 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6" (OuterVolumeSpecName: "kube-api-access-crpr6") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "kube-api-access-crpr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.079976 5018 scope.go:117] "RemoveContainer" containerID="9ee6212a2b63f905b073c82ddf2d4de39d20df1ab5e4d3bc12599b3c148d3400" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.080153 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2" (OuterVolumeSpecName: "kube-api-access-jrdp2") pod "3d48b611-d8d9-48a6-9179-d58ffa16ac18" (UID: "3d48b611-d8d9-48a6-9179-d58ffa16ac18"). InnerVolumeSpecName "kube-api-access-jrdp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.085106 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.089218 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm" (OuterVolumeSpecName: "kube-api-access-7gjhm") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "kube-api-access-7gjhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.105155 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-559f8b448b-zgvh5"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.107814 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq" (OuterVolumeSpecName: "kube-api-access-k6rwq") pod "8fb6f85b-5866-46bf-a826-0a22243318c1" (UID: "8fb6f85b-5866-46bf-a826-0a22243318c1"). InnerVolumeSpecName "kube-api-access-k6rwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.115112 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts" (OuterVolumeSpecName: "scripts") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.116038 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.117144 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.134556 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173182 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173219 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173258 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173296 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173352 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173383 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173421 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173510 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmhn6\" (UniqueName: \"kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173550 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173606 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cst47\" (UniqueName: \"kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173650 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173675 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173756 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173825 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173851 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgt26\" (UniqueName: \"kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173873 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173917 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173939 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173966 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd\") pod \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\" (UID: \"af09a285-4bbe-4a1a-9fa2-9aa1ec047328\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.173991 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts\") pod \"827969fb-cda2-4793-8e73-51b2d1159caa\" (UID: \"827969fb-cda2-4793-8e73-51b2d1159caa\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.174014 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs\") pod \"eba7274f-a215-4784-a540-06013b44a29c\" (UID: \"eba7274f-a215-4784-a540-06013b44a29c\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175015 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6rwq\" (UniqueName: \"kubernetes.io/projected/8fb6f85b-5866-46bf-a826-0a22243318c1-kube-api-access-k6rwq\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175029 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrdp2\" (UniqueName: \"kubernetes.io/projected/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-api-access-jrdp2\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175043 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crpr6\" (UniqueName: \"kubernetes.io/projected/07df2031-64c4-4e6b-b70a-831edefc4468-kube-api-access-crpr6\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175053 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175062 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175074 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07df2031-64c4-4e6b-b70a-831edefc4468-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175084 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gjhm\" (UniqueName: \"kubernetes.io/projected/2a17b000-a726-4072-8b96-04f317ff7fb0-kube-api-access-7gjhm\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.175095 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.176026 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs" (OuterVolumeSpecName: "logs") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.188764 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.189236 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.189700 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.197909 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6" (OuterVolumeSpecName: "kube-api-access-lmhn6") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "kube-api-access-lmhn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.200450 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts" (OuterVolumeSpecName: "scripts") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.204161 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47" (OuterVolumeSpecName: "kube-api-access-cst47") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "kube-api-access-cst47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.208109 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.214037 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.221404 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af09a285-4bbe-4a1a-9fa2-9aa1ec047328","Type":"ContainerDied","Data":"62ae37e1c397e59bdf1a805b4423f2e0bf8bf212741bb9d0d6f8fc36b878d0a7"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.221460 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.221695 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts" (OuterVolumeSpecName: "scripts") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.223962 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.224009 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"05de60b4-2823-4420-b590-2e31ac05b0fe","Type":"ContainerDied","Data":"cea5974d59471076cd0f7a6319db0c799042549cdb4bd3313bd3bf7b52ab9d78"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.225495 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7fe75171-60d0-48df-a61a-c0e828f23f66","Type":"ContainerDied","Data":"09267f3128f593fc994473c067db631355597be81a0f9fc87a429b619699ca40"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.226080 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.230011 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.234114 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26" (OuterVolumeSpecName: "kube-api-access-cgt26") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "kube-api-access-cgt26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.243111 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-576557876d-fngj4" event={"ID":"eba7274f-a215-4784-a540-06013b44a29c","Type":"ContainerDied","Data":"d5d5f6908760c77f469d24ae4f12e61f5f9796e2bc3aa135771af4990ad3b26b"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.243133 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-576557876d-fngj4" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.246843 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" event={"ID":"642159cf-6607-450f-84b9-f8ac137d173d","Type":"ContainerDied","Data":"8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.246892 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a38e304d8e4c5f5b9e5ab8ff9d1576484bb470ba85bd8be321f6c3b048ab4cc" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.257097 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb7c669fd-grgkd" event={"ID":"e3c0a334-9cd6-479b-a2c4-c55063d993a2","Type":"ContainerDied","Data":"7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.257144 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a9e9d9a56e2f5ec1c54c9c86536e6c3c89b1cfe354c6e0ea3c387f2450dc95f" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.259471 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"827969fb-cda2-4793-8e73-51b2d1159caa","Type":"ContainerDied","Data":"38a87e8b8e86e3dd91c7651f3591d41a64c55b22d38f22a130a6f83b2a0b8979"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.259585 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.261828 5018 generic.go:334] "Generic (PLEG): container finished" podID="28d51170-15aa-4edd-9223-ebe057d805b0" containerID="7f1e1ceaf3476f4daf09a417720793400035922c5e45d6ecc1cf736b24eb05d0" exitCode=1 Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.261876 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh033c-account-delete-9nfhv" event={"ID":"28d51170-15aa-4edd-9223-ebe057d805b0","Type":"ContainerDied","Data":"7f1e1ceaf3476f4daf09a417720793400035922c5e45d6ecc1cf736b24eb05d0"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.261896 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh033c-account-delete-9nfhv" event={"ID":"28d51170-15aa-4edd-9223-ebe057d805b0","Type":"ContainerStarted","Data":"d1a0c1238390a2ded4ca6d6ab2cd214e7b1e77eb81cddce94069d8a1b121c834"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.268573 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"941d01f1-c45d-49c3-923b-d685ad3a2f4b","Type":"ContainerDied","Data":"85ec1a5eb8758d65e8e0f323dfef6a54fe81f358c97e49d594c9364bb909b70c"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.268627 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ec1a5eb8758d65e8e0f323dfef6a54fe81f358c97e49d594c9364bb909b70c" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272577 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican3632-account-delete-sgfbk" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272598 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77bf8919-6a28-4eae-8c2c-24d287c5815e","Type":"ContainerDied","Data":"769a9ec21e4cbdd1dabaa7ffa6a86f902f0f3cdeb31c4045eedf45220331d074"} Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272636 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="769a9ec21e4cbdd1dabaa7ffa6a86f902f0f3cdeb31c4045eedf45220331d074" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272761 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272580 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutrond3e5-account-delete-4mrl8" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272812 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5b8df9c49b-fpwlm" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.272909 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.278908 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle\") pod \"05de60b4-2823-4420-b590-2e31ac05b0fe\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279089 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs\") pod \"05de60b4-2823-4420-b590-2e31ac05b0fe\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279123 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data\") pod \"7fe75171-60d0-48df-a61a-c0e828f23f66\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs\") pod \"05de60b4-2823-4420-b590-2e31ac05b0fe\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279556 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs" (OuterVolumeSpecName: "logs") pod "05de60b4-2823-4420-b590-2e31ac05b0fe" (UID: "05de60b4-2823-4420-b590-2e31ac05b0fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279732 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crkkh\" (UniqueName: \"kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh\") pod \"7fe75171-60d0-48df-a61a-c0e828f23f66\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279794 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle\") pod \"7fe75171-60d0-48df-a61a-c0e828f23f66\" (UID: \"7fe75171-60d0-48df-a61a-c0e828f23f66\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279851 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhd9f\" (UniqueName: \"kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f\") pod \"05de60b4-2823-4420-b590-2e31ac05b0fe\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.279868 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data\") pod \"05de60b4-2823-4420-b590-2e31ac05b0fe\" (UID: \"05de60b4-2823-4420-b590-2e31ac05b0fe\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280337 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280352 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmhn6\" (UniqueName: \"kubernetes.io/projected/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-kube-api-access-lmhn6\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280363 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/827969fb-cda2-4793-8e73-51b2d1159caa-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280371 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cst47\" (UniqueName: \"kubernetes.io/projected/eba7274f-a215-4784-a540-06013b44a29c-kube-api-access-cst47\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280380 5018 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280389 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05de60b4-2823-4420-b590-2e31ac05b0fe-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280397 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgt26\" (UniqueName: \"kubernetes.io/projected/827969fb-cda2-4793-8e73-51b2d1159caa-kube-api-access-cgt26\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280405 5018 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280413 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280422 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280431 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.280441 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.280504 5018 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.280552 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data podName:29481fb2-99c9-41f0-b650-a971fa5ff28b nodeName:}" failed. No retries permitted until 2025-10-14 09:02:36.280536655 +0000 UTC m=+7972.864583282 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data") pod "rabbitmq-server-0" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b") : configmap "rabbitmq-config-data" not found Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.320851 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh" (OuterVolumeSpecName: "kube-api-access-crkkh") pod "7fe75171-60d0-48df-a61a-c0e828f23f66" (UID: "7fe75171-60d0-48df-a61a-c0e828f23f66"). InnerVolumeSpecName "kube-api-access-crkkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.322644 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f" (OuterVolumeSpecName: "kube-api-access-qhd9f") pod "05de60b4-2823-4420-b590-2e31ac05b0fe" (UID: "05de60b4-2823-4420-b590-2e31ac05b0fe"). InnerVolumeSpecName "kube-api-access-qhd9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.326930 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.340867 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "3d48b611-d8d9-48a6-9179-d58ffa16ac18" (UID: "3d48b611-d8d9-48a6-9179-d58ffa16ac18"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.344375 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.348387 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.353899 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.353953 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.379868 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.382123 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhd9f\" (UniqueName: \"kubernetes.io/projected/05de60b4-2823-4420-b590-2e31ac05b0fe-kube-api-access-qhd9f\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.382158 5018 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.382168 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.382178 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.382188 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crkkh\" (UniqueName: \"kubernetes.io/projected/7fe75171-60d0-48df-a61a-c0e828f23f66-kube-api-access-crkkh\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.382243 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.382287 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:36.382272697 +0000 UTC m=+7972.966319324 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.382807 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.382844 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:36.382834453 +0000 UTC m=+7972.966881080 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.383174 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:28 crc kubenswrapper[5018]: E1014 09:02:28.383224 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:36.383215614 +0000 UTC m=+7972.967262241 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.394331 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fe75171-60d0-48df-a61a-c0e828f23f66" (UID: "7fe75171-60d0-48df-a61a-c0e828f23f66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.425253 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d48b611-d8d9-48a6-9179-d58ffa16ac18" (UID: "3d48b611-d8d9-48a6-9179-d58ffa16ac18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.451032 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05de60b4-2823-4420-b590-2e31ac05b0fe" (UID: "05de60b4-2823-4420-b590-2e31ac05b0fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.452729 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74b99bc44-qsr8p" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.457540 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data" (OuterVolumeSpecName: "config-data") pod "05de60b4-2823-4420-b590-2e31ac05b0fe" (UID: "05de60b4-2823-4420-b590-2e31ac05b0fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.497278 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.497541 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.497550 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.497558 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.499945 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.504194 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data" (OuterVolumeSpecName: "config-data") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.516004 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data" (OuterVolumeSpecName: "config-data") pod "7fe75171-60d0-48df-a61a-c0e828f23f66" (UID: "7fe75171-60d0-48df-a61a-c0e828f23f66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.553842 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "3d48b611-d8d9-48a6-9179-d58ffa16ac18" (UID: "3d48b611-d8d9-48a6-9179-d58ffa16ac18"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.583795 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data" (OuterVolumeSpecName: "config-data") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.600239 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fe75171-60d0-48df-a61a-c0e828f23f66-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.600265 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.600275 5018 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.600286 5018 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3d48b611-d8d9-48a6-9179-d58ffa16ac18-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.600296 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.623686 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" path="/var/lib/kubelet/pods/46a6263a-f328-43d0-8fcd-fb3610e88c30/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.624355 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" path="/var/lib/kubelet/pods/521da8cb-e7fb-42b0-aa61-ee86ec329fad/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.626289 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.628161 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" path="/var/lib/kubelet/pods/6b5c5d38-56a1-447d-96f4-a8427ff3df82/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.629346 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" path="/var/lib/kubelet/pods/8a5a6c5b-9410-4fde-98c4-112a882b366c/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.629591 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data" (OuterVolumeSpecName: "config-data") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.630215 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" path="/var/lib/kubelet/pods/ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.632666 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" path="/var/lib/kubelet/pods/fc347af0-75b9-4175-bc5e-70f3078e35c6/volumes" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.682423 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "05de60b4-2823-4420-b590-2e31ac05b0fe" (UID: "05de60b4-2823-4420-b590-2e31ac05b0fe"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.686379 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.689764 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.702231 5018 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/05de60b4-2823-4420-b590-2e31ac05b0fe-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.702264 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.702279 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.702292 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.702304 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.708852 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.709055 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.712769 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.716538 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "eba7274f-a215-4784-a540-06013b44a29c" (UID: "eba7274f-a215-4784-a540-06013b44a29c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.721880 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "07df2031-64c4-4e6b-b70a-831edefc4468" (UID: "07df2031-64c4-4e6b-b70a-831edefc4468"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.722872 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.728645 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data" (OuterVolumeSpecName: "config-data") pod "af09a285-4bbe-4a1a-9fa2-9aa1ec047328" (UID: "af09a285-4bbe-4a1a-9fa2-9aa1ec047328"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.731928 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data" (OuterVolumeSpecName: "config-data") pod "827969fb-cda2-4793-8e73-51b2d1159caa" (UID: "827969fb-cda2-4793-8e73-51b2d1159caa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.733606 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2a17b000-a726-4072-8b96-04f317ff7fb0" (UID: "2a17b000-a726-4072-8b96-04f317ff7fb0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.760589 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.761768 5018 scope.go:117] "RemoveContainer" containerID="ac50df93bbc18e55844ca557b4aa55078662999d23f67f0d58048a5477b2f20c" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803647 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803688 5018 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803701 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803712 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a17b000-a726-4072-8b96-04f317ff7fb0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803724 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07df2031-64c4-4e6b-b70a-831edefc4468-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803734 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/827969fb-cda2-4793-8e73-51b2d1159caa-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803743 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af09a285-4bbe-4a1a-9fa2-9aa1ec047328-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803756 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.803783 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eba7274f-a215-4784-a540-06013b44a29c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.849284 5018 scope.go:117] "RemoveContainer" containerID="3c0372fbfcb7d77786b28550cb757bcf2d36775c128fc8d61c2eeed2bb4a27a4" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.851435 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.877600 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.894148 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.904833 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle\") pod \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.904904 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt2lf\" (UniqueName: \"kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf\") pod \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.905000 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data\") pod \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.905051 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config\") pod \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.905218 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs\") pod \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\" (UID: \"941d01f1-c45d-49c3-923b-d685ad3a2f4b\") " Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.906683 5018 scope.go:117] "RemoveContainer" containerID="d43f95354a677e6c3aafe9f165dbe59e283b3bddc5f2df6c04584560150ca6b8" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.907372 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "941d01f1-c45d-49c3-923b-d685ad3a2f4b" (UID: "941d01f1-c45d-49c3-923b-d685ad3a2f4b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.907399 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data" (OuterVolumeSpecName: "config-data") pod "941d01f1-c45d-49c3-923b-d685ad3a2f4b" (UID: "941d01f1-c45d-49c3-923b-d685ad3a2f4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.923061 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf" (OuterVolumeSpecName: "kube-api-access-mt2lf") pod "941d01f1-c45d-49c3-923b-d685ad3a2f4b" (UID: "941d01f1-c45d-49c3-923b-d685ad3a2f4b"). InnerVolumeSpecName "kube-api-access-mt2lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.946585 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "941d01f1-c45d-49c3-923b-d685ad3a2f4b" (UID: "941d01f1-c45d-49c3-923b-d685ad3a2f4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.956911 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.977579 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.977866 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.978475 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.979785 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:28 crc kubenswrapper[5018]: I1014 09:02:28.998945 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.004729 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutrond3e5-account-delete-4mrl8"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.006944 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w87n2\" (UniqueName: \"kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.006995 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007014 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007032 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmvf7\" (UniqueName: \"kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007072 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007099 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007208 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007229 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54qjw\" (UniqueName: \"kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007282 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007308 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007337 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007359 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007406 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007429 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007451 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle\") pod \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\" (UID: \"e3c0a334-9cd6-479b-a2c4-c55063d993a2\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007479 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007517 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007542 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs\") pod \"642159cf-6607-450f-84b9-f8ac137d173d\" (UID: \"642159cf-6607-450f-84b9-f8ac137d173d\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.007560 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs\") pod \"77bf8919-6a28-4eae-8c2c-24d287c5815e\" (UID: \"77bf8919-6a28-4eae-8c2c-24d287c5815e\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.008015 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.008033 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt2lf\" (UniqueName: \"kubernetes.io/projected/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kube-api-access-mt2lf\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.008047 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.008058 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/941d01f1-c45d-49c3-923b-d685ad3a2f4b-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.008490 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs" (OuterVolumeSpecName: "logs") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.012508 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.014787 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2" (OuterVolumeSpecName: "kube-api-access-w87n2") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "kube-api-access-w87n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.015762 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs" (OuterVolumeSpecName: "logs") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.021733 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "941d01f1-c45d-49c3-923b-d685ad3a2f4b" (UID: "941d01f1-c45d-49c3-923b-d685ad3a2f4b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.023205 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.024380 5018 scope.go:117] "RemoveContainer" containerID="ba05b131d9cc289cc0f7d05bb2b92055b3494f32ba43fa2269cb65376bf6bbb9" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.024489 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.024522 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.024826 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican3632-account-delete-sgfbk"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.050939 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.057032 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw" (OuterVolumeSpecName: "kube-api-access-54qjw") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "kube-api-access-54qjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.077930 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7" (OuterVolumeSpecName: "kube-api-access-gmvf7") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "kube-api-access-gmvf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.083662 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.088692 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.099789 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.114310 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hffn\" (UniqueName: \"kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn\") pod \"5876dcf2-c564-45ac-bf38-bebf49e5661f\" (UID: \"5876dcf2-c564-45ac-bf38-bebf49e5661f\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.114561 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfcb8\" (UniqueName: \"kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8\") pod \"7f55eb6e-96bb-4019-83a9-10ba9f455ecc\" (UID: \"7f55eb6e-96bb-4019-83a9-10ba9f455ecc\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.114661 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k2m7\" (UniqueName: \"kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7\") pod \"90c10727-d872-4046-876c-39475c9d8623\" (UID: \"90c10727-d872-4046-876c-39475c9d8623\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115209 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115235 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115250 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115264 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bf8919-6a28-4eae-8c2c-24d287c5815e-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115276 5018 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/941d01f1-c45d-49c3-923b-d685ad3a2f4b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115287 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w87n2\" (UniqueName: \"kubernetes.io/projected/77bf8919-6a28-4eae-8c2c-24d287c5815e-kube-api-access-w87n2\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115300 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmvf7\" (UniqueName: \"kubernetes.io/projected/642159cf-6607-450f-84b9-f8ac137d173d-kube-api-access-gmvf7\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115313 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54qjw\" (UniqueName: \"kubernetes.io/projected/e3c0a334-9cd6-479b-a2c4-c55063d993a2-kube-api-access-54qjw\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115325 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.115335 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3c0a334-9cd6-479b-a2c4-c55063d993a2-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.117012 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.121223 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7" (OuterVolumeSpecName: "kube-api-access-8k2m7") pod "90c10727-d872-4046-876c-39475c9d8623" (UID: "90c10727-d872-4046-876c-39475c9d8623"). InnerVolumeSpecName "kube-api-access-8k2m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.121387 5018 scope.go:117] "RemoveContainer" containerID="82fa2b51a89e84f109bc39c3a296ac0d8bc72f1de856de23122ba2cda3338fdd" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.129237 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.138067 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8" (OuterVolumeSpecName: "kube-api-access-wfcb8") pod "7f55eb6e-96bb-4019-83a9-10ba9f455ecc" (UID: "7f55eb6e-96bb-4019-83a9-10ba9f455ecc"). InnerVolumeSpecName "kube-api-access-wfcb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.144125 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data" (OuterVolumeSpecName: "config-data") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.166278 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn" (OuterVolumeSpecName: "kube-api-access-2hffn") pod "5876dcf2-c564-45ac-bf38-bebf49e5661f" (UID: "5876dcf2-c564-45ac-bf38-bebf49e5661f"). InnerVolumeSpecName "kube-api-access-2hffn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.196702 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.214739 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218404 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218457 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfcb8\" (UniqueName: \"kubernetes.io/projected/7f55eb6e-96bb-4019-83a9-10ba9f455ecc-kube-api-access-wfcb8\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218469 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218480 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k2m7\" (UniqueName: \"kubernetes.io/projected/90c10727-d872-4046-876c-39475c9d8623-kube-api-access-8k2m7\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218489 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hffn\" (UniqueName: \"kubernetes.io/projected/5876dcf2-c564-45ac-bf38-bebf49e5661f-kube-api-access-2hffn\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.218497 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.219076 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-576557876d-fngj4"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.220346 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.225909 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data" (OuterVolumeSpecName: "config-data") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.229738 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data" (OuterVolumeSpecName: "config-data") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.241192 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.246442 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.247685 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e3c0a334-9cd6-479b-a2c4-c55063d993a2" (UID: "e3c0a334-9cd6-479b-a2c4-c55063d993a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.251744 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "642159cf-6607-450f-84b9-f8ac137d173d" (UID: "642159cf-6607-450f-84b9-f8ac137d173d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.252936 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.276024 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "77bf8919-6a28-4eae-8c2c-24d287c5815e" (UID: "77bf8919-6a28-4eae-8c2c-24d287c5815e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.276074 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.310136 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heatd90b-account-delete-sh6hm" event={"ID":"90c10727-d872-4046-876c-39475c9d8623","Type":"ContainerDied","Data":"cfe6a6bea837f6350f884c109dd8566794111b3168f22cf5979fecca71b15957"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.310276 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heatd90b-account-delete-sh6hm" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.312886 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh033c-account-delete-9nfhv" event={"ID":"28d51170-15aa-4edd-9223-ebe057d805b0","Type":"ContainerDied","Data":"d1a0c1238390a2ded4ca6d6ab2cd214e7b1e77eb81cddce94069d8a1b121c834"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.312925 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a0c1238390a2ded4ca6d6ab2cd214e7b1e77eb81cddce94069d8a1b121c834" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.315068 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4d8e88d0-a4e0-4206-bce3-998742afbdc5/ovn-northd/0.log" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.315097 5018 generic.go:334] "Generic (PLEG): container finished" podID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" exitCode=139 Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.315140 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerDied","Data":"d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.315161 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4d8e88d0-a4e0-4206-bce3-998742afbdc5","Type":"ContainerDied","Data":"85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.315170 5018 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ba60c7e76b5099b2413f062a83a73e1da91aff9040deffd8b91627c961d958" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.321572 5018 generic.go:334] "Generic (PLEG): container finished" podID="1383035e-278d-425c-9062-f76183fcdb31" containerID="589d1bdd70b59fd98fe13f906d2c0e56fa29052d702a35f68536bb9f4b51c629" exitCode=0 Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.321768 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerDied","Data":"589d1bdd70b59fd98fe13f906d2c0e56fa29052d702a35f68536bb9f4b51c629"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322338 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322362 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322375 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77bf8919-6a28-4eae-8c2c-24d287c5815e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322448 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322461 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322474 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c0a334-9cd6-479b-a2c4-c55063d993a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.322484 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642159cf-6607-450f-84b9-f8ac137d173d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.324585 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.339573 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement85f4-account-delete-lx2cs" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.341299 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement85f4-account-delete-lx2cs" event={"ID":"5876dcf2-c564-45ac-bf38-bebf49e5661f","Type":"ContainerDied","Data":"98061215156f5ec989c15434fba1483e8d6c14becbf30b9346f581c4159322af"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.346401 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.348371 5018 generic.go:334] "Generic (PLEG): container finished" podID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" containerID="3b364f7b933d0c59e076d99185b783f5158d326517a316e15f35f3ba0599def5" exitCode=0 Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.348510 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5479495b9c-w7q9b" event={"ID":"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46","Type":"ContainerDied","Data":"3b364f7b933d0c59e076d99185b783f5158d326517a316e15f35f3ba0599def5"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.352320 5018 generic.go:334] "Generic (PLEG): container finished" podID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerID="353c1ff5f1490343dd0dfdf4922745f628e70fdbb923936bb2cb5ed954c5db54" exitCode=0 Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.352373 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerDied","Data":"353c1ff5f1490343dd0dfdf4922745f628e70fdbb923936bb2cb5ed954c5db54"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.353332 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5b8df9c49b-fpwlm"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.354397 5018 generic.go:334] "Generic (PLEG): container finished" podID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerID="efb355b1080edb66f752f3f04a2f8a2c9ade67592342b2f5278487f5bddf2b1a" exitCode=0 Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.354444 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerDied","Data":"efb355b1080edb66f752f3f04a2f8a2c9ade67592342b2f5278487f5bddf2b1a"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.357182 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.358246 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.358396 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.359787 5018 scope.go:117] "RemoveContainer" containerID="087e0937d73febf55e6d268d5a2a924ae960ed5b8f8408710ab27cc2646a8605" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.360453 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76cf75d58b-v66bz" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.360519 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance72c6-account-delete-27sjg" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.361431 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb7c669fd-grgkd" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.364352 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance72c6-account-delete-27sjg" event={"ID":"7f55eb6e-96bb-4019-83a9-10ba9f455ecc","Type":"ContainerDied","Data":"3c33d9583bae5f2c62f8c334a794707b142bb53b5e990bfcdba248d32775bc34"} Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.376884 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.380866 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4d8e88d0-a4e0-4206-bce3-998742afbdc5/ovn-northd/0.log" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.380934 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.386741 5018 scope.go:117] "RemoveContainer" containerID="298b04d7b493d03a2ba6b0a665cc5d3190695959aea3d6fec443b33a4c7ab1c7" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.389514 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.390971 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.397244 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.414753 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.423287 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement85f4-account-delete-lx2cs"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.423446 5018 scope.go:117] "RemoveContainer" containerID="68ece5d157ffc84bb5586ef4fa04f0d496bf0dbee2e3cc8cbcd837a81d450889" Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.424930 5018 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: secret "barbican-worker-config-data" not found Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.424987 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:37.424970304 +0000 UTC m=+7974.009016931 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-worker-config-data" not found Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.425052 5018 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.425120 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data podName:6b967f1d-3f1f-414f-87b9-c5e0ee989e23 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:37.425102598 +0000 UTC m=+7974.009149215 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data") pod "barbican-worker-854795b57-c4dnm" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23") : secret "barbican-config-data" not found Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.430867 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.466985 5018 scope.go:117] "RemoveContainer" containerID="ef53fbe1eb381437837cbf0ca10ca7d3c3b0550ab75c3531ca2b8207894fbe27" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.473783 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heatd90b-account-delete-sh6hm"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.491170 5018 scope.go:117] "RemoveContainer" containerID="65b14e1c027a8cae67975b1fce4a4009ca6aec0752d8342f09ee93d4ac090487" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.491287 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.509742 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.525912 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.525979 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.525999 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.526029 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.526078 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.526101 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbkmx\" (UniqueName: \"kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx\") pod \"28d51170-15aa-4edd-9223-ebe057d805b0\" (UID: \"28d51170-15aa-4edd-9223-ebe057d805b0\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.528337 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529458 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529693 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529712 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info" (OuterVolumeSpecName: "pod-info") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529736 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529759 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529781 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529840 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529855 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529895 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529916 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529934 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529959 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529974 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.529992 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kr8g\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530012 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530030 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wqnq\" (UniqueName: \"kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530050 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530073 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530096 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530122 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530140 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir\") pod \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\" (UID: \"4d8e88d0-a4e0-4206-bce3-998742afbdc5\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530164 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzq5c\" (UniqueName: \"kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.530187 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs\") pod \"1383035e-278d-425c-9062-f76183fcdb31\" (UID: \"1383035e-278d-425c-9062-f76183fcdb31\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.531102 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets" (OuterVolumeSpecName: "secrets") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.531888 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.532240 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535025 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535091 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44c35a1d-a43f-4503-a378-4149e63f8e25-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535108 5018 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535659 5018 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535680 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1383035e-278d-425c-9062-f76183fcdb31-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.535564 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.536212 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.536343 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx" (OuterVolumeSpecName: "kube-api-access-hbkmx") pod "28d51170-15aa-4edd-9223-ebe057d805b0" (UID: "28d51170-15aa-4edd-9223-ebe057d805b0"). InnerVolumeSpecName "kube-api-access-hbkmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.536590 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.536900 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config" (OuterVolumeSpecName: "config") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.537997 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts" (OuterVolumeSpecName: "scripts") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.538277 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.538542 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.539119 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.539709 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.541719 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c" (OuterVolumeSpecName: "kube-api-access-fzq5c") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "kube-api-access-fzq5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.547854 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-76cf75d58b-v66bz"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.548074 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.548713 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq" (OuterVolumeSpecName: "kube-api-access-5wqnq") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "kube-api-access-5wqnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.550136 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g" (OuterVolumeSpecName: "kube-api-access-9kr8g") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "kube-api-access-9kr8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.564362 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data" (OuterVolumeSpecName: "config-data") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.564786 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.568356 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.569128 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9" (OuterVolumeSpecName: "mysql-db") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.576994 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5bb7c669fd-grgkd"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.581775 5018 scope.go:117] "RemoveContainer" containerID="43c51dfbbf6f193db78196f33d43664caf8c79a70ac23cfea9d3cc6fa3799171" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.584750 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.588118 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461 podName:44c35a1d-a43f-4503-a378-4149e63f8e25 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:30.088093591 +0000 UTC m=+7966.672140208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "persistence" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.588798 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.592581 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.593748 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.600738 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance72c6-account-delete-27sjg"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.618543 5018 scope.go:117] "RemoveContainer" containerID="c337d9341511431c4a0f5f57944ece931c9e13a68e5f6796f87487029392bc65" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.621593 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.625718 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.637602 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbkmx\" (UniqueName: \"kubernetes.io/projected/28d51170-15aa-4edd-9223-ebe057d805b0-kube-api-access-hbkmx\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.637876 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") on node \"crc\" " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.637977 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638129 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638230 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638313 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638451 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638178 5018 scope.go:117] "RemoveContainer" containerID="89c2c3c1c29bf6244a053cff8a3b895dd672b6f514f9287e56db1eeba83bd95e" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638550 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638697 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kr8g\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-kube-api-access-9kr8g\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638711 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638721 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wqnq\" (UniqueName: \"kubernetes.io/projected/4d8e88d0-a4e0-4206-bce3-998742afbdc5-kube-api-access-5wqnq\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638729 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638739 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638748 5018 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1383035e-278d-425c-9062-f76183fcdb31-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638757 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d8e88d0-a4e0-4206-bce3-998742afbdc5-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638765 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d8e88d0-a4e0-4206-bce3-998742afbdc5-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638773 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzq5c\" (UniqueName: \"kubernetes.io/projected/1383035e-278d-425c-9062-f76183fcdb31-kube-api-access-fzq5c\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.638782 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44c35a1d-a43f-4503-a378-4149e63f8e25-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.641524 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "1383035e-278d-425c-9062-f76183fcdb31" (UID: "1383035e-278d-425c-9062-f76183fcdb31"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.646646 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf" (OuterVolumeSpecName: "server-conf") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.659997 5018 scope.go:117] "RemoveContainer" containerID="b10bff320cd1aef3f540117705e3e341d22490e36c13290cfba7eec50e8f6a71" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.665239 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4d8e88d0-a4e0-4206-bce3-998742afbdc5" (UID: "4d8e88d0-a4e0-4206-bce3-998742afbdc5"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.667276 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.667397 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9") on node "crc" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.680978 5018 scope.go:117] "RemoveContainer" containerID="ae59eb68f544fafc2469fbd82a6797652eaa3a4b90fe15cf3bd03e230ec1f737" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.692725 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.718136 5018 scope.go:117] "RemoveContainer" containerID="677b825788e48fd7761329b03c95970c16bed163be1fd67aade0b8effcd2bbcc" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.740688 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.740767 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.740859 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.741011 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.741086 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.741400 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.741429 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.741488 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l45ft\" (UniqueName: \"kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft\") pod \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\" (UID: \"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46\") " Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.742149 5018 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1383035e-278d-425c-9062-f76183fcdb31-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.742167 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44c35a1d-a43f-4503-a378-4149e63f8e25-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.742177 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b84e6dd-adf5-43a3-b3ea-d551cc2ffbf9\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.742187 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44c35a1d-a43f-4503-a378-4149e63f8e25-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.742197 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d8e88d0-a4e0-4206-bce3-998742afbdc5-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.743943 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.745389 5018 scope.go:117] "RemoveContainer" containerID="99764dec4ca9c95b5f6b94a2065141c0caf1d883b5b4424ac566e1a0775432f3" Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.747103 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.748947 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.749317 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.749444 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts" (OuterVolumeSpecName: "scripts") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.750434 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:02:29 crc kubenswrapper[5018]: E1014 09:02:29.750578 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.752716 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft" (OuterVolumeSpecName: "kube-api-access-l45ft") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "kube-api-access-l45ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.763443 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data" (OuterVolumeSpecName: "config-data") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.765040 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-559f8b448b-zgvh5" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.1.60:8080/healthcheck\": dial tcp 10.217.1.60:8080: i/o timeout" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.765151 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-559f8b448b-zgvh5" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.1.60:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.780677 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.786282 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.811024 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" (UID: "abaeb119-88c5-44b4-8ee5-dc31d1e5cb46"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846824 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846856 5018 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846865 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846875 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l45ft\" (UniqueName: \"kubernetes.io/projected/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-kube-api-access-l45ft\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846883 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846892 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846901 5018 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.846908 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:29 crc kubenswrapper[5018]: I1014 09:02:29.870281 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.049919 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050310 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050417 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050442 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4kbv\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050472 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050684 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050730 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050768 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.050822 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.051499 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.051562 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf\") pod \"29481fb2-99c9-41f0-b650-a971fa5ff28b\" (UID: \"29481fb2-99c9-41f0-b650-a971fa5ff28b\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.052110 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.052337 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.053180 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.054008 5018 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.054026 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.054038 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.056770 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.059233 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv" (OuterVolumeSpecName: "kube-api-access-j4kbv") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "kube-api-access-j4kbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.060349 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.074904 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989" (OuterVolumeSpecName: "persistence") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.082235 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info" (OuterVolumeSpecName: "pod-info") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.093043 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data" (OuterVolumeSpecName: "config-data") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.136014 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf" (OuterVolumeSpecName: "server-conf") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.154725 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") pod \"44c35a1d-a43f-4503-a378-4149e63f8e25\" (UID: \"44c35a1d-a43f-4503-a378-4149e63f8e25\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155102 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155118 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4kbv\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-kube-api-access-j4kbv\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155130 5018 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29481fb2-99c9-41f0-b650-a971fa5ff28b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155139 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155147 5018 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29481fb2-99c9-41f0-b650-a971fa5ff28b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155166 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") on node \"crc\" " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.155175 5018 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29481fb2-99c9-41f0-b650-a971fa5ff28b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: E1014 09:02:30.161522 5018 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4849b9e9_47c2_49bc_aa83_c4a0e7e16bec.slice/crio-84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32a1c03f_068c_4278_b239_480514f4b60d.slice/crio-f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4849b9e9_47c2_49bc_aa83_c4a0e7e16bec.slice/crio-conmon-84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.169386 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461" (OuterVolumeSpecName: "persistence") pod "44c35a1d-a43f-4503-a378-4149e63f8e25" (UID: "44c35a1d-a43f-4503-a378-4149e63f8e25"). InnerVolumeSpecName "pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.183105 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.183276 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989") on node "crc" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.186561 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "29481fb2-99c9-41f0-b650-a971fa5ff28b" (UID: "29481fb2-99c9-41f0-b650-a971fa5ff28b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.257541 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") on node \"crc\" " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.257582 5018 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29481fb2-99c9-41f0-b650-a971fa5ff28b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.257598 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-84fe5d0f-c3a4-417f-8556-2beac7e2b989\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.320329 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.320718 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461") on node "crc" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.359996 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-070a40b5-1186-4d4e-ac78-2d7e3a00b461\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.370863 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44c35a1d-a43f-4503-a378-4149e63f8e25","Type":"ContainerDied","Data":"6ac8491691b7c1d72c212c80aa65eb7ffd3f143eb7c09d519fbe409cc726bef0"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.370906 5018 scope.go:117] "RemoveContainer" containerID="353c1ff5f1490343dd0dfdf4922745f628e70fdbb923936bb2cb5ed954c5db54" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.371084 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.376985 5018 generic.go:334] "Generic (PLEG): container finished" podID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerID="74d5736c20baee1bfbfb8ea1b988e45aea425c479ba0be5662af48c4044d6a13" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.377065 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerDied","Data":"74d5736c20baee1bfbfb8ea1b988e45aea425c479ba0be5662af48c4044d6a13"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.383599 5018 generic.go:334] "Generic (PLEG): container finished" podID="32a1c03f-068c-4278-b239-480514f4b60d" containerID="f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.383650 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerDied","Data":"f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.387011 5018 generic.go:334] "Generic (PLEG): container finished" podID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerID="645fd93c4abe65ac53978d4f7400e542a575a47bf0b11107c5f6b7d78f3348dd" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.387067 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerDied","Data":"645fd93c4abe65ac53978d4f7400e542a575a47bf0b11107c5f6b7d78f3348dd"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.390143 5018 generic.go:334] "Generic (PLEG): container finished" podID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.390209 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8678903d-8a9e-466b-a43c-3aaed9c5e3a5","Type":"ContainerDied","Data":"bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.393317 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29481fb2-99c9-41f0-b650-a971fa5ff28b","Type":"ContainerDied","Data":"546f19dde3cba5be5adacabfec549dd44cba41c39fcbebee003e9c99b62e53b9"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.393471 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.396204 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5479495b9c-w7q9b" event={"ID":"abaeb119-88c5-44b4-8ee5-dc31d1e5cb46","Type":"ContainerDied","Data":"966129542e044847487d9d6857b9b661f73d0f0e9338a64e570c36e0b08f54c4"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.396287 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5479495b9c-w7q9b" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.399225 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1383035e-278d-425c-9062-f76183fcdb31","Type":"ContainerDied","Data":"5bd6d3322f32483a18d6f7509f368d72cea2754b64b07b54604a283d01808f05"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.399293 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403281 5018 generic.go:334] "Generic (PLEG): container finished" podID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerID="84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403311 5018 generic.go:334] "Generic (PLEG): container finished" podID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerID="4d5980364138e871981e0c2a309c64682bdee2ba69842dd2a7eb988420399d1e" exitCode=0 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403352 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerDied","Data":"84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403379 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403436 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh033c-account-delete-9nfhv" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.403383 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerDied","Data":"4d5980364138e871981e0c2a309c64682bdee2ba69842dd2a7eb988420399d1e"} Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.448056 5018 scope.go:117] "RemoveContainer" containerID="f4d99450585994209d83212376bd59bf0d8a61b2535a8ac45829db06d6c64d20" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.451173 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.472199 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.482893 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.491456 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.497187 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.508735 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5479495b9c-w7q9b"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.517987 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.524156 5018 scope.go:117] "RemoveContainer" containerID="efb355b1080edb66f752f3f04a2f8a2c9ade67592342b2f5278487f5bddf2b1a" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.524312 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.529413 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.534869 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.539317 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.543865 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh033c-account-delete-9nfhv"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.544766 5018 scope.go:117] "RemoveContainer" containerID="abe3f15398c9cdd459afab86842e412ed3e64bbadd7f82be725fd3caaad3b66c" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.568684 5018 scope.go:117] "RemoveContainer" containerID="3b364f7b933d0c59e076d99185b783f5158d326517a316e15f35f3ba0599def5" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.613609 5018 scope.go:117] "RemoveContainer" containerID="589d1bdd70b59fd98fe13f906d2c0e56fa29052d702a35f68536bb9f4b51c629" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.615563 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" path="/var/lib/kubelet/pods/05de60b4-2823-4420-b590-2e31ac05b0fe/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.616175 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" path="/var/lib/kubelet/pods/07df2031-64c4-4e6b-b70a-831edefc4468/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.616834 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1383035e-278d-425c-9062-f76183fcdb31" path="/var/lib/kubelet/pods/1383035e-278d-425c-9062-f76183fcdb31/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.617834 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d51170-15aa-4edd-9223-ebe057d805b0" path="/var/lib/kubelet/pods/28d51170-15aa-4edd-9223-ebe057d805b0/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.618430 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" path="/var/lib/kubelet/pods/29481fb2-99c9-41f0-b650-a971fa5ff28b/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.619478 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" path="/var/lib/kubelet/pods/2a17b000-a726-4072-8b96-04f317ff7fb0/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.620004 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" path="/var/lib/kubelet/pods/3d48b611-d8d9-48a6-9179-d58ffa16ac18/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.620917 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" path="/var/lib/kubelet/pods/44c35a1d-a43f-4503-a378-4149e63f8e25/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.621489 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" path="/var/lib/kubelet/pods/4d8e88d0-a4e0-4206-bce3-998742afbdc5/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.621967 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5876dcf2-c564-45ac-bf38-bebf49e5661f" path="/var/lib/kubelet/pods/5876dcf2-c564-45ac-bf38-bebf49e5661f/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.622917 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642159cf-6607-450f-84b9-f8ac137d173d" path="/var/lib/kubelet/pods/642159cf-6607-450f-84b9-f8ac137d173d/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.623516 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" path="/var/lib/kubelet/pods/77bf8919-6a28-4eae-8c2c-24d287c5815e/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.623994 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f55eb6e-96bb-4019-83a9-10ba9f455ecc" path="/var/lib/kubelet/pods/7f55eb6e-96bb-4019-83a9-10ba9f455ecc/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.624413 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" path="/var/lib/kubelet/pods/7fe75171-60d0-48df-a61a-c0e828f23f66/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.625334 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" path="/var/lib/kubelet/pods/827969fb-cda2-4793-8e73-51b2d1159caa/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.625961 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fb6f85b-5866-46bf-a826-0a22243318c1" path="/var/lib/kubelet/pods/8fb6f85b-5866-46bf-a826-0a22243318c1/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.626493 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c10727-d872-4046-876c-39475c9d8623" path="/var/lib/kubelet/pods/90c10727-d872-4046-876c-39475c9d8623/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.627548 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" path="/var/lib/kubelet/pods/941d01f1-c45d-49c3-923b-d685ad3a2f4b/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.628038 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" path="/var/lib/kubelet/pods/abaeb119-88c5-44b4-8ee5-dc31d1e5cb46/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.628543 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" path="/var/lib/kubelet/pods/af09a285-4bbe-4a1a-9fa2-9aa1ec047328/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.629661 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" path="/var/lib/kubelet/pods/e3c0a334-9cd6-479b-a2c4-c55063d993a2/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.630149 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab7a625-e4b8-4038-b9a7-ce1f7af786bf" path="/var/lib/kubelet/pods/eab7a625-e4b8-4038-b9a7-ce1f7af786bf/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.630715 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba7274f-a215-4784-a540-06013b44a29c" path="/var/lib/kubelet/pods/eba7274f-a215-4784-a540-06013b44a29c/volumes" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.637062 5018 scope.go:117] "RemoveContainer" containerID="97f38af704283bb352ed02f09332ddce7ba2726b12ca8ffa9339d06c33f07de1" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.692015 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.878882 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs\") pod \"32a1c03f-068c-4278-b239-480514f4b60d\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.878945 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom\") pod \"32a1c03f-068c-4278-b239-480514f4b60d\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.878976 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle\") pod \"32a1c03f-068c-4278-b239-480514f4b60d\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.878997 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt7w6\" (UniqueName: \"kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6\") pod \"32a1c03f-068c-4278-b239-480514f4b60d\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.879032 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data\") pod \"32a1c03f-068c-4278-b239-480514f4b60d\" (UID: \"32a1c03f-068c-4278-b239-480514f4b60d\") " Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.882005 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs" (OuterVolumeSpecName: "logs") pod "32a1c03f-068c-4278-b239-480514f4b60d" (UID: "32a1c03f-068c-4278-b239-480514f4b60d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.889636 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "32a1c03f-068c-4278-b239-480514f4b60d" (UID: "32a1c03f-068c-4278-b239-480514f4b60d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.893548 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6" (OuterVolumeSpecName: "kube-api-access-zt7w6") pod "32a1c03f-068c-4278-b239-480514f4b60d" (UID: "32a1c03f-068c-4278-b239-480514f4b60d"). InnerVolumeSpecName "kube-api-access-zt7w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.899080 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.899304 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="35063fdb-abe2-4d3a-8833-8d62162a1a01" containerName="adoption" containerID="cri-o://9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c" gracePeriod=30 Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.920602 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32a1c03f-068c-4278-b239-480514f4b60d" (UID: "32a1c03f-068c-4278-b239-480514f4b60d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.956231 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data" (OuterVolumeSpecName: "config-data") pod "32a1c03f-068c-4278-b239-480514f4b60d" (UID: "32a1c03f-068c-4278-b239-480514f4b60d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.980464 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32a1c03f-068c-4278-b239-480514f4b60d-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.980847 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.980858 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.980867 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt7w6\" (UniqueName: \"kubernetes.io/projected/32a1c03f-068c-4278-b239-480514f4b60d-kube-api-access-zt7w6\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.980876 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32a1c03f-068c-4278-b239-480514f4b60d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:30 crc kubenswrapper[5018]: I1014 09:02:30.997350 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.020836 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.032715 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.059946 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.190839 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.191091 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="43c4ab03-dd63-4001-9a55-833b15bc11fc" containerName="adoption" containerID="cri-o://fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18" gracePeriod=30 Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192757 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192784 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192807 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192823 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192845 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs\") pod \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.192863 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v4ms\" (UniqueName: \"kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193259 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle\") pod \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193351 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193370 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193531 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193562 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle\") pod \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193582 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data\") pod \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193608 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data\") pod \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193634 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193659 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom\") pod \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193682 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwlbr\" (UniqueName: \"kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193707 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwwc8\" (UniqueName: \"kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8\") pod \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\" (UID: \"6b967f1d-3f1f-414f-87b9-c5e0ee989e23\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193731 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193747 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle\") pod \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\" (UID: \"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193777 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll8h7\" (UniqueName: \"kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7\") pod \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\" (UID: \"8678903d-8a9e-466b-a43c-3aaed9c5e3a5\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.193823 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs\") pod \"b1598984-ef08-45c2-a286-fd527ca31a7d\" (UID: \"b1598984-ef08-45c2-a286-fd527ca31a7d\") " Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.196984 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs" (OuterVolumeSpecName: "logs") pod "6b967f1d-3f1f-414f-87b9-c5e0ee989e23" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.215424 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6b967f1d-3f1f-414f-87b9-c5e0ee989e23" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.216937 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.217509 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7" (OuterVolumeSpecName: "kube-api-access-ll8h7") pod "8678903d-8a9e-466b-a43c-3aaed9c5e3a5" (UID: "8678903d-8a9e-466b-a43c-3aaed9c5e3a5"). InnerVolumeSpecName "kube-api-access-ll8h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.218522 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8678903d-8a9e-466b-a43c-3aaed9c5e3a5" (UID: "8678903d-8a9e-466b-a43c-3aaed9c5e3a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.219768 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts" (OuterVolumeSpecName: "scripts") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.219909 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms" (OuterVolumeSpecName: "kube-api-access-5v4ms") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "kube-api-access-5v4ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.221137 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr" (OuterVolumeSpecName: "kube-api-access-vwlbr") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "kube-api-access-vwlbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.227670 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8" (OuterVolumeSpecName: "kube-api-access-kwwc8") pod "6b967f1d-3f1f-414f-87b9-c5e0ee989e23" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23"). InnerVolumeSpecName "kube-api-access-kwwc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.251267 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data" (OuterVolumeSpecName: "config-data") pod "8678903d-8a9e-466b-a43c-3aaed9c5e3a5" (UID: "8678903d-8a9e-466b-a43c-3aaed9c5e3a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.259187 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b967f1d-3f1f-414f-87b9-c5e0ee989e23" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.262733 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.271545 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data" (OuterVolumeSpecName: "config-data") pod "6b967f1d-3f1f-414f-87b9-c5e0ee989e23" (UID: "6b967f1d-3f1f-414f-87b9-c5e0ee989e23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.278204 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.282612 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.283964 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.290840 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295029 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295054 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295063 5018 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295072 5018 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295082 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295091 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v4ms\" (UniqueName: \"kubernetes.io/projected/b1598984-ef08-45c2-a286-fd527ca31a7d-kube-api-access-5v4ms\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295101 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295110 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295118 5018 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295127 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295140 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295155 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295163 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295171 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwlbr\" (UniqueName: \"kubernetes.io/projected/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-kube-api-access-vwlbr\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295180 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwwc8\" (UniqueName: \"kubernetes.io/projected/6b967f1d-3f1f-414f-87b9-c5e0ee989e23-kube-api-access-kwwc8\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295189 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.295198 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll8h7\" (UniqueName: \"kubernetes.io/projected/8678903d-8a9e-466b-a43c-3aaed9c5e3a5-kube-api-access-ll8h7\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.299933 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config" (OuterVolumeSpecName: "config") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.302935 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b1598984-ef08-45c2-a286-fd527ca31a7d" (UID: "b1598984-ef08-45c2-a286-fd527ca31a7d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.314320 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data" (OuterVolumeSpecName: "config-data") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.327325 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" (UID: "4849b9e9-47c2-49bc-aa83-c4a0e7e16bec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.396357 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.396389 5018 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.396400 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.396410 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b1598984-ef08-45c2-a286-fd527ca31a7d-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.414395 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8678903d-8a9e-466b-a43c-3aaed9c5e3a5","Type":"ContainerDied","Data":"9af70c0d3318bdaa60014bb4acc2befa4ff36520ed41f356ac8dc928b223bb9f"} Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.414441 5018 scope.go:117] "RemoveContainer" containerID="bea5685d91dc4ede119f0794c2bad707ffb146d1034cd201f8a7fc681d8d6789" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.414478 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.420521 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" event={"ID":"32a1c03f-068c-4278-b239-480514f4b60d","Type":"ContainerDied","Data":"90a6b0d8e9a7e5c57565c31b12754efe38f9de81afbd9e4c85903fa8a6636544"} Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.420611 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dcf959744-4fkwc" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.436826 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-854795b57-c4dnm" event={"ID":"6b967f1d-3f1f-414f-87b9-c5e0ee989e23","Type":"ContainerDied","Data":"1562d007f4b144548257c422a2911771c538a8901d3797f5334540d7bb00b957"} Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.436992 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-854795b57-c4dnm" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.440449 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4849b9e9-47c2-49bc-aa83-c4a0e7e16bec","Type":"ContainerDied","Data":"b12d8b4a9229f30afca9bd014771ef78b595311ae1791a5f1e757fde4319d886"} Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.440546 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.457667 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.460348 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd4dcdf-tjqrh" event={"ID":"b1598984-ef08-45c2-a286-fd527ca31a7d","Type":"ContainerDied","Data":"fa6fa0270e513ee2cd3cd6af8513862162036f0f4c655f09b390f49042708c52"} Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.460434 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd4dcdf-tjqrh" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.470850 5018 scope.go:117] "RemoveContainer" containerID="f7405b97a2cd94f5e13b32f5176354ad1c06d0bcc1272b576bd196d24b0d1eb1" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.471501 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.471530 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.148:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.487657 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.500164 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5dcf959744-4fkwc"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.504795 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.505027 5018 scope.go:117] "RemoveContainer" containerID="0e6d9ef3c618844329be04c22cf07c96416cade5d7551e947ded7ff89d6b027d" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.512879 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.518713 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.523260 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-854795b57-c4dnm"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.529225 5018 scope.go:117] "RemoveContainer" containerID="74d5736c20baee1bfbfb8ea1b988e45aea425c479ba0be5662af48c4044d6a13" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.529561 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.535554 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69dd4dcdf-tjqrh"] Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.548543 5018 scope.go:117] "RemoveContainer" containerID="e8b14272735d2836638396ff83b7ead777ae1c594b519c1e3e4bf3dc34b6cf8b" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.573208 5018 scope.go:117] "RemoveContainer" containerID="84530d98987f93364f3bc5a750b6e83d9187a5b7565f3f43d1dd05f231a65096" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.589779 5018 scope.go:117] "RemoveContainer" containerID="4d5980364138e871981e0c2a309c64682bdee2ba69842dd2a7eb988420399d1e" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.617119 5018 scope.go:117] "RemoveContainer" containerID="a7585a2f1c37963c5dbf63572362ad5333d952de5f762503637a079285e5bd70" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.636782 5018 scope.go:117] "RemoveContainer" containerID="2d821f28ca619425ae5555a780634062a928277adbe2e77a11f5825e3ed63b9d" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.656161 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.1.68:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.660526 5018 scope.go:117] "RemoveContainer" containerID="b153baeb0241dacb32e954510d08cddc263a2d3ee9e254846604179d4e462655" Oct 14 09:02:31 crc kubenswrapper[5018]: I1014 09:02:31.683612 5018 scope.go:117] "RemoveContainer" containerID="645fd93c4abe65ac53978d4f7400e542a575a47bf0b11107c5f6b7d78f3348dd" Oct 14 09:02:32 crc kubenswrapper[5018]: I1014 09:02:32.619347 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a1c03f-068c-4278-b239-480514f4b60d" path="/var/lib/kubelet/pods/32a1c03f-068c-4278-b239-480514f4b60d/volumes" Oct 14 09:02:32 crc kubenswrapper[5018]: I1014 09:02:32.620693 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" path="/var/lib/kubelet/pods/4849b9e9-47c2-49bc-aa83-c4a0e7e16bec/volumes" Oct 14 09:02:32 crc kubenswrapper[5018]: I1014 09:02:32.621563 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" path="/var/lib/kubelet/pods/6b967f1d-3f1f-414f-87b9-c5e0ee989e23/volumes" Oct 14 09:02:32 crc kubenswrapper[5018]: I1014 09:02:32.622774 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" path="/var/lib/kubelet/pods/8678903d-8a9e-466b-a43c-3aaed9c5e3a5/volumes" Oct 14 09:02:32 crc kubenswrapper[5018]: I1014 09:02:32.623331 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" path="/var/lib/kubelet/pods/b1598984-ef08-45c2-a286-fd527ca31a7d/volumes" Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407123 5018 configmap.go:193] Couldn't get configMap openstack/horizon-scripts: configmap "horizon-scripts" not found Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407464 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:52.407445112 +0000 UTC m=+7988.991491749 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-scripts" not found Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407176 5018 configmap.go:193] Couldn't get configMap openstack/horizon-config-data: configmap "horizon-config-data" not found Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407176 5018 secret.go:188] Couldn't get secret openstack/horizon: secret "horizon" not found Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407696 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:52.407651898 +0000 UTC m=+7988.991698565 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : configmap "horizon-config-data" not found Oct 14 09:02:36 crc kubenswrapper[5018]: E1014 09:02:36.407739 5018 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key podName:2cd4a8e5-6715-4830-a0df-134660996996 nodeName:}" failed. No retries permitted until 2025-10-14 09:02:52.4077212 +0000 UTC m=+7988.991767977 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "horizon-secret-key" (UniqueName: "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key") pod "horizon-74b99bc44-qsr8p" (UID: "2cd4a8e5-6715-4830-a0df-134660996996") : secret "horizon" not found Oct 14 09:02:37 crc kubenswrapper[5018]: I1014 09:02:37.605738 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:02:37 crc kubenswrapper[5018]: E1014 09:02:37.606081 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:02:38 crc kubenswrapper[5018]: E1014 09:02:38.340125 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:38 crc kubenswrapper[5018]: E1014 09:02:38.341705 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:38 crc kubenswrapper[5018]: E1014 09:02:38.342866 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:38 crc kubenswrapper[5018]: E1014 09:02:38.342915 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:02:38 crc kubenswrapper[5018]: I1014 09:02:38.452988 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74b99bc44-qsr8p" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Oct 14 09:02:48 crc kubenswrapper[5018]: E1014 09:02:48.339409 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:48 crc kubenswrapper[5018]: E1014 09:02:48.344892 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:48 crc kubenswrapper[5018]: E1014 09:02:48.350081 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:48 crc kubenswrapper[5018]: E1014 09:02:48.350161 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:02:48 crc kubenswrapper[5018]: I1014 09:02:48.456616 5018 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74b99bc44-qsr8p" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Oct 14 09:02:48 crc kubenswrapper[5018]: I1014 09:02:48.456781 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 09:02:48 crc kubenswrapper[5018]: I1014 09:02:48.608746 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:02:48 crc kubenswrapper[5018]: E1014 09:02:48.609026 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:02:51 crc kubenswrapper[5018]: I1014 09:02:51.721180 5018 generic.go:334] "Generic (PLEG): container finished" podID="2cd4a8e5-6715-4830-a0df-134660996996" containerID="dcb994f2a796078d866b79a0686477e41a0d979852b18ac2243d743ecb903f27" exitCode=137 Oct 14 09:02:51 crc kubenswrapper[5018]: I1014 09:02:51.721836 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerDied","Data":"dcb994f2a796078d866b79a0686477e41a0d979852b18ac2243d743ecb903f27"} Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.012504 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.120862 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121008 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121030 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121062 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121090 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121118 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.121161 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44lm5\" (UniqueName: \"kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5\") pod \"2cd4a8e5-6715-4830-a0df-134660996996\" (UID: \"2cd4a8e5-6715-4830-a0df-134660996996\") " Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.122128 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs" (OuterVolumeSpecName: "logs") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.126515 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.136080 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5" (OuterVolumeSpecName: "kube-api-access-44lm5") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "kube-api-access-44lm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.145353 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts" (OuterVolumeSpecName: "scripts") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.145720 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data" (OuterVolumeSpecName: "config-data") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.156579 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.178249 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "2cd4a8e5-6715-4830-a0df-134660996996" (UID: "2cd4a8e5-6715-4830-a0df-134660996996"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222837 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222879 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222895 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222908 5018 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2cd4a8e5-6715-4830-a0df-134660996996-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222920 5018 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cd4a8e5-6715-4830-a0df-134660996996-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222933 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cd4a8e5-6715-4830-a0df-134660996996-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.222945 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44lm5\" (UniqueName: \"kubernetes.io/projected/2cd4a8e5-6715-4830-a0df-134660996996-kube-api-access-44lm5\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.733723 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b99bc44-qsr8p" event={"ID":"2cd4a8e5-6715-4830-a0df-134660996996","Type":"ContainerDied","Data":"fc66057e4599459dbcba725f4871e513169b148b99ab3a0f13948e5fb286a3f2"} Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.733780 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b99bc44-qsr8p" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.733802 5018 scope.go:117] "RemoveContainer" containerID="392ee2ce4169eab5c3bfa508967090781008c6bb4b91e91bc63269bdd41562f3" Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.763769 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.771888 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74b99bc44-qsr8p"] Oct 14 09:02:52 crc kubenswrapper[5018]: I1014 09:02:52.911449 5018 scope.go:117] "RemoveContainer" containerID="dcb994f2a796078d866b79a0686477e41a0d979852b18ac2243d743ecb903f27" Oct 14 09:02:54 crc kubenswrapper[5018]: I1014 09:02:54.630432 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cd4a8e5-6715-4830-a0df-134660996996" path="/var/lib/kubelet/pods/2cd4a8e5-6715-4830-a0df-134660996996/volumes" Oct 14 09:02:54 crc kubenswrapper[5018]: I1014 09:02:54.702967 5018 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podb67bc93c-1c02-47fa-b2a5-57560256e89b"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podb67bc93c-1c02-47fa-b2a5-57560256e89b] : Timed out while waiting for systemd to remove kubepods-besteffort-podb67bc93c_1c02_47fa_b2a5_57560256e89b.slice" Oct 14 09:02:54 crc kubenswrapper[5018]: I1014 09:02:54.705688 5018 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod9703a145-2c14-4a43-a767-e6def8fe6063"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod9703a145-2c14-4a43-a767-e6def8fe6063] : Timed out while waiting for systemd to remove kubepods-besteffort-pod9703a145_2c14_4a43_a767_e6def8fe6063.slice" Oct 14 09:02:57 crc kubenswrapper[5018]: I1014 09:02:57.909000 5018 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod5f24304f-0a66-4c0f-b8cb-3341f9fb926d"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod5f24304f-0a66-4c0f-b8cb-3341f9fb926d] : Timed out while waiting for systemd to remove kubepods-burstable-pod5f24304f_0a66_4c0f_b8cb_3341f9fb926d.slice" Oct 14 09:02:57 crc kubenswrapper[5018]: E1014 09:02:57.909355 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod5f24304f-0a66-4c0f-b8cb-3341f9fb926d] : unable to destroy cgroup paths for cgroup [kubepods burstable pod5f24304f-0a66-4c0f-b8cb-3341f9fb926d] : Timed out while waiting for systemd to remove kubepods-burstable-pod5f24304f_0a66_4c0f_b8cb_3341f9fb926d.slice" pod="openstack/alertmanager-metric-storage-0" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" Oct 14 09:02:58 crc kubenswrapper[5018]: E1014 09:02:58.340347 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:58 crc kubenswrapper[5018]: E1014 09:02:58.343240 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:58 crc kubenswrapper[5018]: E1014 09:02:58.345186 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:02:58 crc kubenswrapper[5018]: E1014 09:02:58.345243 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:02:58 crc kubenswrapper[5018]: I1014 09:02:58.801575 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:02:58 crc kubenswrapper[5018]: I1014 09:02:58.829529 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:02:58 crc kubenswrapper[5018]: I1014 09:02:58.857121 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:03:00 crc kubenswrapper[5018]: I1014 09:03:00.624995 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" path="/var/lib/kubelet/pods/5f24304f-0a66-4c0f-b8cb-3341f9fb926d/volumes" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.402697 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.507233 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzv95\" (UniqueName: \"kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95\") pod \"35063fdb-abe2-4d3a-8833-8d62162a1a01\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.508020 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") pod \"35063fdb-abe2-4d3a-8833-8d62162a1a01\" (UID: \"35063fdb-abe2-4d3a-8833-8d62162a1a01\") " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.513359 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95" (OuterVolumeSpecName: "kube-api-access-fzv95") pod "35063fdb-abe2-4d3a-8833-8d62162a1a01" (UID: "35063fdb-abe2-4d3a-8833-8d62162a1a01"). InnerVolumeSpecName "kube-api-access-fzv95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.519173 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6" (OuterVolumeSpecName: "mariadb-data") pod "35063fdb-abe2-4d3a-8833-8d62162a1a01" (UID: "35063fdb-abe2-4d3a-8833-8d62162a1a01"). InnerVolumeSpecName "pvc-c8f25d93-1670-421a-804b-fb11581e6ce6". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.593256 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.607590 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:03:01 crc kubenswrapper[5018]: E1014 09:03:01.608199 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.609568 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") on node \"crc\" " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.609685 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzv95\" (UniqueName: \"kubernetes.io/projected/35063fdb-abe2-4d3a-8833-8d62162a1a01-kube-api-access-fzv95\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.640875 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.641083 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c8f25d93-1670-421a-804b-fb11581e6ce6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6") on node "crc" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.710411 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fpkw\" (UniqueName: \"kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw\") pod \"43c4ab03-dd63-4001-9a55-833b15bc11fc\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.710505 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert\") pod \"43c4ab03-dd63-4001-9a55-833b15bc11fc\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.711468 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") pod \"43c4ab03-dd63-4001-9a55-833b15bc11fc\" (UID: \"43c4ab03-dd63-4001-9a55-833b15bc11fc\") " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.711938 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c8f25d93-1670-421a-804b-fb11581e6ce6\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.713428 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw" (OuterVolumeSpecName: "kube-api-access-9fpkw") pod "43c4ab03-dd63-4001-9a55-833b15bc11fc" (UID: "43c4ab03-dd63-4001-9a55-833b15bc11fc"). InnerVolumeSpecName "kube-api-access-9fpkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.713698 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "43c4ab03-dd63-4001-9a55-833b15bc11fc" (UID: "43c4ab03-dd63-4001-9a55-833b15bc11fc"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.724422 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474" (OuterVolumeSpecName: "ovn-data") pod "43c4ab03-dd63-4001-9a55-833b15bc11fc" (UID: "43c4ab03-dd63-4001-9a55-833b15bc11fc"). InnerVolumeSpecName "pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.813807 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fpkw\" (UniqueName: \"kubernetes.io/projected/43c4ab03-dd63-4001-9a55-833b15bc11fc-kube-api-access-9fpkw\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.813873 5018 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/43c4ab03-dd63-4001-9a55-833b15bc11fc-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.813945 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") on node \"crc\" " Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.846192 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.846487 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474") on node "crc" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.846935 5018 generic.go:334] "Generic (PLEG): container finished" podID="43c4ab03-dd63-4001-9a55-833b15bc11fc" containerID="fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18" exitCode=137 Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.847020 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"43c4ab03-dd63-4001-9a55-833b15bc11fc","Type":"ContainerDied","Data":"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18"} Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.847073 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"43c4ab03-dd63-4001-9a55-833b15bc11fc","Type":"ContainerDied","Data":"7cc99a0f83af0d084556ee8f65b92f6ed60d9a41a2cd3497f97f3da4f22ac714"} Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.847105 5018 scope.go:117] "RemoveContainer" containerID="fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.847289 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.849844 5018 generic.go:334] "Generic (PLEG): container finished" podID="35063fdb-abe2-4d3a-8833-8d62162a1a01" containerID="9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c" exitCode=137 Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.849882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"35063fdb-abe2-4d3a-8833-8d62162a1a01","Type":"ContainerDied","Data":"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c"} Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.849908 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"35063fdb-abe2-4d3a-8833-8d62162a1a01","Type":"ContainerDied","Data":"a9b712927c6c2a994d8f40700c7dd0ed2da3263737850e387fd081507abaa976"} Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.849911 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.928234 5018 scope.go:117] "RemoveContainer" containerID="fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.928405 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.929159 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9a81c3d-5f58-4ce2-b68c-de4643cdf474\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:01 crc kubenswrapper[5018]: E1014 09:03:01.932700 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18\": container with ID starting with fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18 not found: ID does not exist" containerID="fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.932759 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18"} err="failed to get container status \"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18\": rpc error: code = NotFound desc = could not find container \"fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18\": container with ID starting with fc8da0a4eced9f9c5793ae05848579597a5621a88fbda95834dff5116e6f7c18 not found: ID does not exist" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.932812 5018 scope.go:117] "RemoveContainer" containerID="9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.951887 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.956576 5018 scope.go:117] "RemoveContainer" containerID="9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c" Oct 14 09:03:01 crc kubenswrapper[5018]: E1014 09:03:01.957083 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c\": container with ID starting with 9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c not found: ID does not exist" containerID="9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.957119 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c"} err="failed to get container status \"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c\": rpc error: code = NotFound desc = could not find container \"9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c\": container with ID starting with 9759e45c53ef6af9c303bb9fa567f520d5479f7fabe2ec5a43783d8752cbb22c not found: ID does not exist" Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.964863 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:03:01 crc kubenswrapper[5018]: I1014 09:03:01.972755 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:03:02 crc kubenswrapper[5018]: I1014 09:03:02.622665 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35063fdb-abe2-4d3a-8833-8d62162a1a01" path="/var/lib/kubelet/pods/35063fdb-abe2-4d3a-8833-8d62162a1a01/volumes" Oct 14 09:03:02 crc kubenswrapper[5018]: I1014 09:03:02.624150 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c4ab03-dd63-4001-9a55-833b15bc11fc" path="/var/lib/kubelet/pods/43c4ab03-dd63-4001-9a55-833b15bc11fc/volumes" Oct 14 09:03:08 crc kubenswrapper[5018]: E1014 09:03:08.339512 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:08 crc kubenswrapper[5018]: E1014 09:03:08.342062 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:08 crc kubenswrapper[5018]: E1014 09:03:08.344085 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:08 crc kubenswrapper[5018]: E1014 09:03:08.344293 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:03:14 crc kubenswrapper[5018]: I1014 09:03:14.609635 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:03:14 crc kubenswrapper[5018]: E1014 09:03:14.610289 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.722514 5018 scope.go:117] "RemoveContainer" containerID="43bfd079138bb3c2e9ec5326544cf4acf7e96fc023b1eadc3c147d3b1433cbb2" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.804579 5018 scope.go:117] "RemoveContainer" containerID="b5014c9bfac19f9acded35ea492bf1ac02255d3437039504d0647dddd1fdb140" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.844204 5018 scope.go:117] "RemoveContainer" containerID="d58eb2f2536f3484883eea2d8e6c8fd61155a9efbf660b0648f5a63ed92f951b" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.867080 5018 scope.go:117] "RemoveContainer" containerID="b85abd63af53151ec3ccb50a2dbc7b82688f80e4c0d1741e99d29d0486758ea1" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.901149 5018 scope.go:117] "RemoveContainer" containerID="e75f3722943cc4d4b297dc5162439acb46bfe1e62a4291fce3f2582331b91793" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.945292 5018 scope.go:117] "RemoveContainer" containerID="9b40efcd93a346933415044024e90ede001495fcfd8fb0320885bab4c64fda7d" Oct 14 09:03:16 crc kubenswrapper[5018]: I1014 09:03:16.979970 5018 scope.go:117] "RemoveContainer" containerID="40fb8b68d8bb04c1d344475d4f00e45548e0b4c28983fcb23f3b012d39221907" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.018653 5018 scope.go:117] "RemoveContainer" containerID="815430d7601a06039490b1146f9c8d9aafce21bf0b6517a403372225de85d41d" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.050674 5018 scope.go:117] "RemoveContainer" containerID="cc378e4ea02d43f477ec8203ef137c674e7055a3b57a90d0bac877f9fdd800da" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.080868 5018 scope.go:117] "RemoveContainer" containerID="f3bfacf0b50c0a9351a97df8cd9a4d4d20587d8f3b2250add56a2512071e49f6" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.095643 5018 scope.go:117] "RemoveContainer" containerID="8e5bc391462c1807435f3141698540e3dbc744434e432fffc72f05920e38ffb9" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.113037 5018 scope.go:117] "RemoveContainer" containerID="ad2b4811966987ea30370b369c8a6b465e5f24ef16c13b9903cbbbc919d7aa05" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.161671 5018 scope.go:117] "RemoveContainer" containerID="377bff2c1f590ed92d30d5e19176b01620ec07050618bc1c81fbd84b01fd16ca" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.183070 5018 scope.go:117] "RemoveContainer" containerID="49e3187ba65a221c522e49e4c2118872d506a95aa0ad3018a57afeb0f6482e51" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.203364 5018 scope.go:117] "RemoveContainer" containerID="e470dea1a4e6ffec6409d94627e8b61dab6f61e9336b8490b9c80edaa9a40879" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.238509 5018 scope.go:117] "RemoveContainer" containerID="c6b8156153448df4ea7ab21d1a8bb3213deefc4aed5b741b5a54c1beb1983b69" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.268305 5018 scope.go:117] "RemoveContainer" containerID="0cf0d1c6030e09c2eda77accb9e964aeb0ff1f0dbf949d67d6d6824cf055ea0a" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.299026 5018 scope.go:117] "RemoveContainer" containerID="652fef03d0063382f0ac1925aeb8888b7617bf80e601c16452dfdee4b721b2c2" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.334999 5018 scope.go:117] "RemoveContainer" containerID="d4b9ecf9551b7d71f24978b3596fb93c674a49dbabf2d694c982d16be6b892f8" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.365300 5018 scope.go:117] "RemoveContainer" containerID="ce612af3722b887703ef66d47d90f1bdbe0b44f623bc74647fe5ba7231238eed" Oct 14 09:03:17 crc kubenswrapper[5018]: I1014 09:03:17.394770 5018 scope.go:117] "RemoveContainer" containerID="8ea98fcf3c27a68c727221403d77e0a9e5e402b4f567534f6bb06689161fc3eb" Oct 14 09:03:18 crc kubenswrapper[5018]: E1014 09:03:18.341183 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:18 crc kubenswrapper[5018]: E1014 09:03:18.345107 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:18 crc kubenswrapper[5018]: E1014 09:03:18.347540 5018 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 14 09:03:18 crc kubenswrapper[5018]: E1014 09:03:18.347673 5018 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.140742 5018 generic.go:334] "Generic (PLEG): container finished" podID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" exitCode=137 Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.141202 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" event={"ID":"214ab07d-ba44-499b-bbad-9a9ec59eeaa3","Type":"ContainerDied","Data":"7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b"} Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.328276 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.412707 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle\") pod \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.412777 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt9x7\" (UniqueName: \"kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7\") pod \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.412840 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom\") pod \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.412931 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data\") pod \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\" (UID: \"214ab07d-ba44-499b-bbad-9a9ec59eeaa3\") " Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.417430 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7" (OuterVolumeSpecName: "kube-api-access-jt9x7") pod "214ab07d-ba44-499b-bbad-9a9ec59eeaa3" (UID: "214ab07d-ba44-499b-bbad-9a9ec59eeaa3"). InnerVolumeSpecName "kube-api-access-jt9x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.417867 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "214ab07d-ba44-499b-bbad-9a9ec59eeaa3" (UID: "214ab07d-ba44-499b-bbad-9a9ec59eeaa3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.432383 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "214ab07d-ba44-499b-bbad-9a9ec59eeaa3" (UID: "214ab07d-ba44-499b-bbad-9a9ec59eeaa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.451871 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data" (OuterVolumeSpecName: "config-data") pod "214ab07d-ba44-499b-bbad-9a9ec59eeaa3" (UID: "214ab07d-ba44-499b-bbad-9a9ec59eeaa3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.513870 5018 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.513914 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.513931 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt9x7\" (UniqueName: \"kubernetes.io/projected/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-kube-api-access-jt9x7\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:23 crc kubenswrapper[5018]: I1014 09:03:23.513942 5018 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214ab07d-ba44-499b-bbad-9a9ec59eeaa3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.155353 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" event={"ID":"214ab07d-ba44-499b-bbad-9a9ec59eeaa3","Type":"ContainerDied","Data":"4c8e3e7942f8d8ca1b6a5a662c86ee3404d18a5b685c656a0a9aea0866777d66"} Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.155418 5018 scope.go:117] "RemoveContainer" containerID="7d696f23b5481a876ee49b64381720a12e6c16ac9fb1ef24fb2fbf73163de10b" Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.155486 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67b5f9ffc9-8mcjc" Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.204309 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.214225 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-67b5f9ffc9-8mcjc"] Oct 14 09:03:24 crc kubenswrapper[5018]: I1014 09:03:24.625162 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" path="/var/lib/kubelet/pods/214ab07d-ba44-499b-bbad-9a9ec59eeaa3/volumes" Oct 14 09:03:26 crc kubenswrapper[5018]: I1014 09:03:26.605490 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:03:26 crc kubenswrapper[5018]: E1014 09:03:26.606130 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:03:38 crc kubenswrapper[5018]: I1014 09:03:38.645922 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:03:38 crc kubenswrapper[5018]: E1014 09:03:38.646877 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:03:50 crc kubenswrapper[5018]: I1014 09:03:50.605808 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:03:50 crc kubenswrapper[5018]: E1014 09:03:50.606976 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.747375 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748424 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748443 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748471 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748484 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748500 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748511 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748529 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="cinder-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748538 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="cinder-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748554 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748564 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748576 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="init" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748586 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="init" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748611 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerName="nova-cell1-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748643 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerName="nova-cell1-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748662 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748671 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748691 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="probe" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748702 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="probe" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748723 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748733 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748747 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="thanos-sidecar" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748757 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="thanos-sidecar" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748779 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748789 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748809 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" containerName="memcached" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748819 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" containerName="memcached" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748868 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="prometheus" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748906 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="prometheus" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748926 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748937 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748958 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.748968 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.748986 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749009 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749022 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c10727-d872-4046-876c-39475c9d8623" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749032 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c10727-d872-4046-876c-39475c9d8623" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749045 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f55eb6e-96bb-4019-83a9-10ba9f455ecc" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749056 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f55eb6e-96bb-4019-83a9-10ba9f455ecc" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749078 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerName="kube-state-metrics" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749088 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerName="kube-state-metrics" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749105 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb6f85b-5866-46bf-a826-0a22243318c1" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749114 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb6f85b-5866-46bf-a826-0a22243318c1" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749133 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-notification-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749144 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-notification-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749177 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749188 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749210 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749222 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749241 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642159cf-6607-450f-84b9-f8ac137d173d" containerName="heat-cfnapi" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749251 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="642159cf-6607-450f-84b9-f8ac137d173d" containerName="heat-cfnapi" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749270 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35063fdb-abe2-4d3a-8833-8d62162a1a01" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749280 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="35063fdb-abe2-4d3a-8833-8d62162a1a01" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749293 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749303 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749315 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749324 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749339 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba7274f-a215-4784-a540-06013b44a29c" containerName="heat-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749349 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba7274f-a215-4784-a540-06013b44a29c" containerName="heat-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749365 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749377 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749390 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-server" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749400 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-server" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749409 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5876dcf2-c564-45ac-bf38-bebf49e5661f" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749419 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5876dcf2-c564-45ac-bf38-bebf49e5661f" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749430 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749440 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749484 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749496 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749516 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749527 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749541 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="sg-core" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749551 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="sg-core" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749569 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" containerName="keystone-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749578 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" containerName="keystone-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749589 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749698 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749753 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="mysql-bootstrap" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749770 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="mysql-bootstrap" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749802 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749814 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749836 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="init-config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749848 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="init-config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749872 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749884 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749901 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerName="nova-cell0-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749949 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerName="nova-cell0-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.749978 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.749990 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750005 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750016 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750031 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-central-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750042 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-central-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750067 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="ovsdbserver-sb" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750078 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="ovsdbserver-sb" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750093 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750106 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750128 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9830a080-9886-4018-b68c-60bda6dd6333" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750155 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9830a080-9886-4018-b68c-60bda6dd6333" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750173 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="setup-container" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750185 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="setup-container" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750199 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750210 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750224 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="init-config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750236 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="init-config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750258 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750270 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750281 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="ovsdbserver-nb" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750292 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="ovsdbserver-nb" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750303 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="dnsmasq-dns" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750313 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="dnsmasq-dns" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750331 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750341 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750361 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750372 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750391 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750404 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750423 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="setup-container" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750435 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="setup-container" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750454 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-evaluator" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750465 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-evaluator" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750484 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d51170-15aa-4edd-9223-ebe057d805b0" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750495 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d51170-15aa-4edd-9223-ebe057d805b0" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750509 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750519 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750536 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab7a625-e4b8-4038-b9a7-ce1f7af786bf" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750546 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab7a625-e4b8-4038-b9a7-ce1f7af786bf" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750561 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c4ab03-dd63-4001-9a55-833b15bc11fc" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750572 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c4ab03-dd63-4001-9a55-833b15bc11fc" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750587 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="alertmanager" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750597 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="alertmanager" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750615 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750649 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750667 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-notifier" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750678 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-notifier" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750699 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750710 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750727 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750738 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750750 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750771 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750783 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750794 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750813 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="mysql-bootstrap" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750824 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="mysql-bootstrap" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750838 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750850 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-api" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750873 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750883 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" Oct 14 09:03:57 crc kubenswrapper[5018]: E1014 09:03:57.750899 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.750909 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751299 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-metadata" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751317 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f55eb6e-96bb-4019-83a9-10ba9f455ecc" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751337 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751349 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751369 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9830a080-9886-4018-b68c-60bda6dd6333" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751385 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751407 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751429 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="1383035e-278d-425c-9062-f76183fcdb31" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751444 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751465 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b967f1d-3f1f-414f-87b9-c5e0ee989e23" containerName="barbican-worker" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751477 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751488 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751506 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751519 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="cinder-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751535 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="prometheus" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751555 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="827969fb-cda2-4793-8e73-51b2d1159caa" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751574 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce8e2f9a-ae4c-4c0a-9062-5b2df44d3aab" containerName="thanos-sidecar" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751589 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751603 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="35063fdb-abe2-4d3a-8833-8d62162a1a01" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751638 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751651 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="sg-core" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751675 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751695 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="05de60b4-2823-4420-b590-2e31ac05b0fe" containerName="nova-metadata-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751709 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="alertmanager" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751726 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="521da8cb-e7fb-42b0-aa61-ee86ec329fad" containerName="proxy-server" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751748 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a17b000-a726-4072-8b96-04f317ff7fb0" containerName="placement-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751762 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751775 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="941d01f1-c45d-49c3-923b-d685ad3a2f4b" containerName="memcached" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751794 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5876dcf2-c564-45ac-bf38-bebf49e5661f" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751816 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="214ab07d-ba44-499b-bbad-9a9ec59eeaa3" containerName="heat-engine" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751831 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f24304f-0a66-4c0f-b8cb-3341f9fb926d" containerName="config-reloader" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751846 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751868 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d8e88d0-a4e0-4206-bce3-998742afbdc5" containerName="ovn-northd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751910 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8678903d-8a9e-466b-a43c-3aaed9c5e3a5" containerName="nova-scheduler-scheduler" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751926 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="openstack-network-exporter" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751948 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-notifier" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751961 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751985 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67bc93c-1c02-47fa-b2a5-57560256e89b" containerName="ovsdbserver-nb" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.751999 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="proxy-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752018 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fb6f85b-5866-46bf-a826-0a22243318c1" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752034 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5c5d38-56a1-447d-96f4-a8427ff3df82" containerName="probe" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752054 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bf8919-6a28-4eae-8c2c-24d287c5815e" containerName="nova-api-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752068 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-central-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752081 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="af09a285-4bbe-4a1a-9fa2-9aa1ec047328" containerName="ceilometer-notification-agent" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752096 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-evaluator" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752115 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1598984-ef08-45c2-a286-fd527ca31a7d" containerName="neutron-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752134 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a6263a-f328-43d0-8fcd-fb3610e88c30" containerName="nova-cell1-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752150 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-httpd" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752165 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="07df2031-64c4-4e6b-b70a-831edefc4468" containerName="cinder-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752178 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5a6c5b-9410-4fde-98c4-112a882b366c" containerName="galera" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752194 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a1c03f-068c-4278-b239-480514f4b60d" containerName="barbican-keystone-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752216 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab7a625-e4b8-4038-b9a7-ce1f7af786bf" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752229 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="642159cf-6607-450f-84b9-f8ac137d173d" containerName="heat-cfnapi" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752244 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe75171-60d0-48df-a61a-c0e828f23f66" containerName="nova-cell0-conductor-conductor" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752256 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55ca4b1-1c7a-49ea-bf68-c32a91de7c6e" containerName="dnsmasq-dns" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752275 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc347af0-75b9-4175-bc5e-70f3078e35c6" containerName="glance-log" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752287 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="4849b9e9-47c2-49bc-aa83-c4a0e7e16bec" containerName="aodh-listener" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752303 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d51170-15aa-4edd-9223-ebe057d805b0" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752319 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c35a1d-a43f-4503-a378-4149e63f8e25" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752334 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c4ab03-dd63-4001-9a55-833b15bc11fc" containerName="adoption" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752351 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c0a334-9cd6-479b-a2c4-c55063d993a2" containerName="barbican-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752372 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d48b611-d8d9-48a6-9179-d58ffa16ac18" containerName="kube-state-metrics" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752390 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba7274f-a215-4784-a540-06013b44a29c" containerName="heat-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752402 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703a145-2c14-4a43-a767-e6def8fe6063" containerName="ovsdbserver-sb" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752416 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c10727-d872-4046-876c-39475c9d8623" containerName="mariadb-account-delete" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752431 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="29481fb2-99c9-41f0-b650-a971fa5ff28b" containerName="rabbitmq" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752451 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cd4a8e5-6715-4830-a0df-134660996996" containerName="horizon" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.752467 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaeb119-88c5-44b4-8ee5-dc31d1e5cb46" containerName="keystone-api" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.754217 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.766757 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.849541 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xtnw\" (UniqueName: \"kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.849611 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.849906 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.951913 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xtnw\" (UniqueName: \"kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.951986 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.952080 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.952651 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.952820 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:57 crc kubenswrapper[5018]: I1014 09:03:57.982074 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xtnw\" (UniqueName: \"kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw\") pod \"redhat-marketplace-bbgln\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:58 crc kubenswrapper[5018]: I1014 09:03:58.097475 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:03:58 crc kubenswrapper[5018]: I1014 09:03:58.580191 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:03:59 crc kubenswrapper[5018]: I1014 09:03:59.522598 5018 generic.go:334] "Generic (PLEG): container finished" podID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerID="70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64" exitCode=0 Oct 14 09:03:59 crc kubenswrapper[5018]: I1014 09:03:59.522754 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerDied","Data":"70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64"} Oct 14 09:03:59 crc kubenswrapper[5018]: I1014 09:03:59.522789 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerStarted","Data":"8e10b651a232f0c1a4f450b50557ad8cda3a6aa72d3e279da535469ac5ce9267"} Oct 14 09:03:59 crc kubenswrapper[5018]: I1014 09:03:59.524894 5018 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:04:01 crc kubenswrapper[5018]: I1014 09:04:01.547579 5018 generic.go:334] "Generic (PLEG): container finished" podID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerID="4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9" exitCode=0 Oct 14 09:04:01 crc kubenswrapper[5018]: I1014 09:04:01.547678 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerDied","Data":"4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9"} Oct 14 09:04:01 crc kubenswrapper[5018]: I1014 09:04:01.605528 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:04:01 crc kubenswrapper[5018]: E1014 09:04:01.605952 5018 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-g7t5z_openshift-machine-config-operator(124fa7dd-4f8d-4d66-9223-f6bcc6df7242)\"" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" Oct 14 09:04:03 crc kubenswrapper[5018]: I1014 09:04:03.575255 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerStarted","Data":"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844"} Oct 14 09:04:03 crc kubenswrapper[5018]: I1014 09:04:03.614366 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bbgln" podStartSLOduration=3.694914535 podStartE2EDuration="6.614345563s" podCreationTimestamp="2025-10-14 09:03:57 +0000 UTC" firstStartedPulling="2025-10-14 09:03:59.524657994 +0000 UTC m=+8056.108704621" lastFinishedPulling="2025-10-14 09:04:02.444088992 +0000 UTC m=+8059.028135649" observedRunningTime="2025-10-14 09:04:03.597931855 +0000 UTC m=+8060.181978502" watchObservedRunningTime="2025-10-14 09:04:03.614345563 +0000 UTC m=+8060.198392190" Oct 14 09:04:08 crc kubenswrapper[5018]: I1014 09:04:08.098570 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:08 crc kubenswrapper[5018]: I1014 09:04:08.098980 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:08 crc kubenswrapper[5018]: I1014 09:04:08.149937 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:08 crc kubenswrapper[5018]: I1014 09:04:08.703082 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:08 crc kubenswrapper[5018]: I1014 09:04:08.762055 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:04:10 crc kubenswrapper[5018]: I1014 09:04:10.646891 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bbgln" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="registry-server" containerID="cri-o://b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844" gracePeriod=2 Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.192247 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.287571 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities\") pod \"705dbb2c-820e-4ec7-8145-7b65afee33d9\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.287989 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xtnw\" (UniqueName: \"kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw\") pod \"705dbb2c-820e-4ec7-8145-7b65afee33d9\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.288037 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content\") pod \"705dbb2c-820e-4ec7-8145-7b65afee33d9\" (UID: \"705dbb2c-820e-4ec7-8145-7b65afee33d9\") " Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.288662 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities" (OuterVolumeSpecName: "utilities") pod "705dbb2c-820e-4ec7-8145-7b65afee33d9" (UID: "705dbb2c-820e-4ec7-8145-7b65afee33d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.304447 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "705dbb2c-820e-4ec7-8145-7b65afee33d9" (UID: "705dbb2c-820e-4ec7-8145-7b65afee33d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.305526 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw" (OuterVolumeSpecName: "kube-api-access-4xtnw") pod "705dbb2c-820e-4ec7-8145-7b65afee33d9" (UID: "705dbb2c-820e-4ec7-8145-7b65afee33d9"). InnerVolumeSpecName "kube-api-access-4xtnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.390000 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xtnw\" (UniqueName: \"kubernetes.io/projected/705dbb2c-820e-4ec7-8145-7b65afee33d9-kube-api-access-4xtnw\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.390422 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.390710 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705dbb2c-820e-4ec7-8145-7b65afee33d9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.658762 5018 generic.go:334] "Generic (PLEG): container finished" podID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerID="b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844" exitCode=0 Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.658802 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerDied","Data":"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844"} Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.658846 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bbgln" event={"ID":"705dbb2c-820e-4ec7-8145-7b65afee33d9","Type":"ContainerDied","Data":"8e10b651a232f0c1a4f450b50557ad8cda3a6aa72d3e279da535469ac5ce9267"} Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.658869 5018 scope.go:117] "RemoveContainer" containerID="b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.658911 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bbgln" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.680559 5018 scope.go:117] "RemoveContainer" containerID="4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.715751 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.721585 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bbgln"] Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.731902 5018 scope.go:117] "RemoveContainer" containerID="70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.759998 5018 scope.go:117] "RemoveContainer" containerID="b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844" Oct 14 09:04:11 crc kubenswrapper[5018]: E1014 09:04:11.760498 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844\": container with ID starting with b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844 not found: ID does not exist" containerID="b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.760565 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844"} err="failed to get container status \"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844\": rpc error: code = NotFound desc = could not find container \"b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844\": container with ID starting with b80956291f9efe570fdf91f9ec4f42f0ded23dac83f79414129f265e7987e844 not found: ID does not exist" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.760606 5018 scope.go:117] "RemoveContainer" containerID="4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9" Oct 14 09:04:11 crc kubenswrapper[5018]: E1014 09:04:11.761141 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9\": container with ID starting with 4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9 not found: ID does not exist" containerID="4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.761169 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9"} err="failed to get container status \"4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9\": rpc error: code = NotFound desc = could not find container \"4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9\": container with ID starting with 4b941db7257ebc2e7109cbf66639d5508b07d143646706f04edf672debdcb1f9 not found: ID does not exist" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.761192 5018 scope.go:117] "RemoveContainer" containerID="70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64" Oct 14 09:04:11 crc kubenswrapper[5018]: E1014 09:04:11.761780 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64\": container with ID starting with 70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64 not found: ID does not exist" containerID="70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64" Oct 14 09:04:11 crc kubenswrapper[5018]: I1014 09:04:11.761824 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64"} err="failed to get container status \"70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64\": rpc error: code = NotFound desc = could not find container \"70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64\": container with ID starting with 70b842068ea14a54a417516fa726784805fafaff05ee2acd4f883b27d5667e64 not found: ID does not exist" Oct 14 09:04:12 crc kubenswrapper[5018]: I1014 09:04:12.618102 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" path="/var/lib/kubelet/pods/705dbb2c-820e-4ec7-8145-7b65afee33d9/volumes" Oct 14 09:04:16 crc kubenswrapper[5018]: I1014 09:04:16.605519 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:04:17 crc kubenswrapper[5018]: I1014 09:04:17.716825 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"b98d0e5c81187860cd3ad149820c384a66b7867dc7a4fbc14fec8624d602899f"} Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.425003 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-945sk/must-gather-vs44p"] Oct 14 09:04:23 crc kubenswrapper[5018]: E1014 09:04:23.425687 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="extract-content" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.425699 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="extract-content" Oct 14 09:04:23 crc kubenswrapper[5018]: E1014 09:04:23.425711 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="extract-utilities" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.425717 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="extract-utilities" Oct 14 09:04:23 crc kubenswrapper[5018]: E1014 09:04:23.425729 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="registry-server" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.425735 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="registry-server" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.425901 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="705dbb2c-820e-4ec7-8145-7b65afee33d9" containerName="registry-server" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.426605 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.428389 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-945sk"/"openshift-service-ca.crt" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.428546 5018 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-945sk"/"kube-root-ca.crt" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.430904 5018 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-945sk"/"default-dockercfg-h8qq9" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.435056 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-945sk/must-gather-vs44p"] Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.488672 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.488727 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7fqr\" (UniqueName: \"kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.590344 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.590433 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7fqr\" (UniqueName: \"kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.590878 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.607200 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7fqr\" (UniqueName: \"kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr\") pod \"must-gather-vs44p\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:23 crc kubenswrapper[5018]: I1014 09:04:23.771073 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:04:24 crc kubenswrapper[5018]: I1014 09:04:24.269531 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-945sk/must-gather-vs44p"] Oct 14 09:04:24 crc kubenswrapper[5018]: I1014 09:04:24.787451 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/must-gather-vs44p" event={"ID":"e38af198-4b74-412f-9f28-fe5ea89a4e67","Type":"ContainerStarted","Data":"c2d2a789a2814e2497b24b5e282866c52141a53f5dc66373d0e79e83310c100c"} Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.619405 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-945sk/crc-debug-k9ckj"] Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.621188 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.713163 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fv7n\" (UniqueName: \"kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.713262 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.814868 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fv7n\" (UniqueName: \"kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.814945 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.815046 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.831450 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/must-gather-vs44p" event={"ID":"e38af198-4b74-412f-9f28-fe5ea89a4e67","Type":"ContainerStarted","Data":"8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6"} Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.831494 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/must-gather-vs44p" event={"ID":"e38af198-4b74-412f-9f28-fe5ea89a4e67","Type":"ContainerStarted","Data":"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989"} Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.852540 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-945sk/must-gather-vs44p" podStartSLOduration=2.5336357830000003 podStartE2EDuration="6.852513087s" podCreationTimestamp="2025-10-14 09:04:23 +0000 UTC" firstStartedPulling="2025-10-14 09:04:24.277177435 +0000 UTC m=+8080.861224102" lastFinishedPulling="2025-10-14 09:04:28.596054739 +0000 UTC m=+8085.180101406" observedRunningTime="2025-10-14 09:04:29.844242011 +0000 UTC m=+8086.428288638" watchObservedRunningTime="2025-10-14 09:04:29.852513087 +0000 UTC m=+8086.436559734" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.862111 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fv7n\" (UniqueName: \"kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n\") pod \"crc-debug-k9ckj\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:29 crc kubenswrapper[5018]: I1014 09:04:29.942422 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:04:30 crc kubenswrapper[5018]: I1014 09:04:30.839857 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/crc-debug-k9ckj" event={"ID":"9da29496-e859-4c94-8bae-1aebbb64b99d","Type":"ContainerStarted","Data":"d1f96db65a031b7e1da30e61d29538a2224b0d420ec2f29ba82f97975f4016c3"} Oct 14 09:04:41 crc kubenswrapper[5018]: I1014 09:04:41.955756 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/crc-debug-k9ckj" event={"ID":"9da29496-e859-4c94-8bae-1aebbb64b99d","Type":"ContainerStarted","Data":"e4d43f76986b202869f8d19da4ac7a9b1b8833ae1ebf8bf7f61c2f892420ca24"} Oct 14 09:04:41 crc kubenswrapper[5018]: I1014 09:04:41.973256 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-945sk/crc-debug-k9ckj" podStartSLOduration=1.553720383 podStartE2EDuration="12.973239986s" podCreationTimestamp="2025-10-14 09:04:29 +0000 UTC" firstStartedPulling="2025-10-14 09:04:29.989833933 +0000 UTC m=+8086.573880560" lastFinishedPulling="2025-10-14 09:04:41.409353526 +0000 UTC m=+8097.993400163" observedRunningTime="2025-10-14 09:04:41.970012484 +0000 UTC m=+8098.554059121" watchObservedRunningTime="2025-10-14 09:04:41.973239986 +0000 UTC m=+8098.557286613" Oct 14 09:05:03 crc kubenswrapper[5018]: I1014 09:05:03.127031 5018 generic.go:334] "Generic (PLEG): container finished" podID="9da29496-e859-4c94-8bae-1aebbb64b99d" containerID="e4d43f76986b202869f8d19da4ac7a9b1b8833ae1ebf8bf7f61c2f892420ca24" exitCode=0 Oct 14 09:05:03 crc kubenswrapper[5018]: I1014 09:05:03.127502 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/crc-debug-k9ckj" event={"ID":"9da29496-e859-4c94-8bae-1aebbb64b99d","Type":"ContainerDied","Data":"e4d43f76986b202869f8d19da4ac7a9b1b8833ae1ebf8bf7f61c2f892420ca24"} Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.201315 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.221158 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-945sk/crc-debug-k9ckj"] Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.225084 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-945sk/crc-debug-k9ckj"] Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.247196 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host\") pod \"9da29496-e859-4c94-8bae-1aebbb64b99d\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.247248 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fv7n\" (UniqueName: \"kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n\") pod \"9da29496-e859-4c94-8bae-1aebbb64b99d\" (UID: \"9da29496-e859-4c94-8bae-1aebbb64b99d\") " Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.247530 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host" (OuterVolumeSpecName: "host") pod "9da29496-e859-4c94-8bae-1aebbb64b99d" (UID: "9da29496-e859-4c94-8bae-1aebbb64b99d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.262102 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n" (OuterVolumeSpecName: "kube-api-access-6fv7n") pod "9da29496-e859-4c94-8bae-1aebbb64b99d" (UID: "9da29496-e859-4c94-8bae-1aebbb64b99d"). InnerVolumeSpecName "kube-api-access-6fv7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.348785 5018 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da29496-e859-4c94-8bae-1aebbb64b99d-host\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.348816 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fv7n\" (UniqueName: \"kubernetes.io/projected/9da29496-e859-4c94-8bae-1aebbb64b99d-kube-api-access-6fv7n\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:04 crc kubenswrapper[5018]: I1014 09:05:04.614827 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9da29496-e859-4c94-8bae-1aebbb64b99d" path="/var/lib/kubelet/pods/9da29496-e859-4c94-8bae-1aebbb64b99d/volumes" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.146716 5018 scope.go:117] "RemoveContainer" containerID="e4d43f76986b202869f8d19da4ac7a9b1b8833ae1ebf8bf7f61c2f892420ca24" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.146810 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-k9ckj" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.430660 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-945sk/crc-debug-9jnqg"] Oct 14 09:05:05 crc kubenswrapper[5018]: E1014 09:05:05.431062 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9da29496-e859-4c94-8bae-1aebbb64b99d" containerName="container-00" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.431081 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="9da29496-e859-4c94-8bae-1aebbb64b99d" containerName="container-00" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.431306 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="9da29496-e859-4c94-8bae-1aebbb64b99d" containerName="container-00" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.431955 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.466729 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzwdg\" (UniqueName: \"kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.466795 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.567868 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.568020 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.568214 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzwdg\" (UniqueName: \"kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.595135 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzwdg\" (UniqueName: \"kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg\") pod \"crc-debug-9jnqg\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:05 crc kubenswrapper[5018]: I1014 09:05:05.746714 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:06 crc kubenswrapper[5018]: I1014 09:05:06.160404 5018 generic.go:334] "Generic (PLEG): container finished" podID="ab39f288-65d7-4f16-86f0-c1d834239898" containerID="45782a9c76038112bc9453c5f7f50c5494d090b2be0f945e32e8f30eaea89545" exitCode=1 Oct 14 09:05:06 crc kubenswrapper[5018]: I1014 09:05:06.160519 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/crc-debug-9jnqg" event={"ID":"ab39f288-65d7-4f16-86f0-c1d834239898","Type":"ContainerDied","Data":"45782a9c76038112bc9453c5f7f50c5494d090b2be0f945e32e8f30eaea89545"} Oct 14 09:05:06 crc kubenswrapper[5018]: I1014 09:05:06.160882 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/crc-debug-9jnqg" event={"ID":"ab39f288-65d7-4f16-86f0-c1d834239898","Type":"ContainerStarted","Data":"5e29ce9681de67e7d1dd8bdbfa072417e522c66eb7b03dc2e5794451634c45f8"} Oct 14 09:05:06 crc kubenswrapper[5018]: I1014 09:05:06.200403 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-945sk/crc-debug-9jnqg"] Oct 14 09:05:06 crc kubenswrapper[5018]: I1014 09:05:06.212701 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-945sk/crc-debug-9jnqg"] Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.246154 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.295103 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host\") pod \"ab39f288-65d7-4f16-86f0-c1d834239898\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.295237 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwdg\" (UniqueName: \"kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg\") pod \"ab39f288-65d7-4f16-86f0-c1d834239898\" (UID: \"ab39f288-65d7-4f16-86f0-c1d834239898\") " Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.295227 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host" (OuterVolumeSpecName: "host") pod "ab39f288-65d7-4f16-86f0-c1d834239898" (UID: "ab39f288-65d7-4f16-86f0-c1d834239898"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.316131 5018 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab39f288-65d7-4f16-86f0-c1d834239898-host\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.318998 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg" (OuterVolumeSpecName: "kube-api-access-nzwdg") pod "ab39f288-65d7-4f16-86f0-c1d834239898" (UID: "ab39f288-65d7-4f16-86f0-c1d834239898"). InnerVolumeSpecName "kube-api-access-nzwdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:07 crc kubenswrapper[5018]: I1014 09:05:07.417535 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwdg\" (UniqueName: \"kubernetes.io/projected/ab39f288-65d7-4f16-86f0-c1d834239898-kube-api-access-nzwdg\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:08 crc kubenswrapper[5018]: I1014 09:05:08.186399 5018 scope.go:117] "RemoveContainer" containerID="45782a9c76038112bc9453c5f7f50c5494d090b2be0f945e32e8f30eaea89545" Oct 14 09:05:08 crc kubenswrapper[5018]: I1014 09:05:08.186458 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/crc-debug-9jnqg" Oct 14 09:05:08 crc kubenswrapper[5018]: I1014 09:05:08.617036 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab39f288-65d7-4f16-86f0-c1d834239898" path="/var/lib/kubelet/pods/ab39f288-65d7-4f16-86f0-c1d834239898/volumes" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.349736 5018 scope.go:117] "RemoveContainer" containerID="0835e3a86cc1bed0b46ea89d60d79c970f2f5d7aa91a959459086c75a2a30e00" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.367635 5018 scope.go:117] "RemoveContainer" containerID="b301967d4689fd683a3b9fe925bc51719be3a31512b8b531cf8d4a04678a438c" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.396104 5018 scope.go:117] "RemoveContainer" containerID="e75da823fed7de7b31f4d43e999ab665655abab1201befd87f9e5a24f0190080" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.417349 5018 scope.go:117] "RemoveContainer" containerID="cf8f965d2cb92a7c4505cb3c64420b7047da837b854170fa012a0781465f8cb5" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.446418 5018 scope.go:117] "RemoveContainer" containerID="77311f9821ca7a858e16455376cf6227d93c7f441b52752f6083c5cada76a306" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.469579 5018 scope.go:117] "RemoveContainer" containerID="2e77272e500e6600d36b4d5079bc8d3c5187606d43f141ea94c262c6476cace8" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.488754 5018 scope.go:117] "RemoveContainer" containerID="5475b3ac1ffb8d1aa80f71e6a95e7eed4b52f74bd503013962bda9fef820ce2f" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.529809 5018 scope.go:117] "RemoveContainer" containerID="10ea6304fb5ac010339e3324db27b0c137c67f17ab90bcea9ecfed1e0cc68529" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.549379 5018 scope.go:117] "RemoveContainer" containerID="00899a309e02428a711c0dbf8cf3c13507e7bba8ba46443809cca45c5e90dcc3" Oct 14 09:05:18 crc kubenswrapper[5018]: I1014 09:05:18.569110 5018 scope.go:117] "RemoveContainer" containerID="d9c7dac13e2dea7e6557db75768e7930258f2cf88772355a4297edfc995e3ef6" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.364821 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bcdada18-a44b-4d39-b4f6-2bc9fef4530c/openstack-network-exporter/0.log" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.520145 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bcdada18-a44b-4d39-b4f6-2bc9fef4530c/ovsdbserver-nb/0.log" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.567794 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_495a63bc-82b8-4a2b-83ab-5aed23f6f493/openstack-network-exporter/0.log" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.695047 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_495a63bc-82b8-4a2b-83ab-5aed23f6f493/ovsdbserver-nb/0.log" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.833408 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/openstack-network-exporter/0.log" Oct 14 09:05:19 crc kubenswrapper[5018]: I1014 09:05:19.900883 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/ovsdbserver-sb/0.log" Oct 14 09:05:20 crc kubenswrapper[5018]: I1014 09:05:20.024718 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_64ed49b9-ade8-4fe4-9c37-ae327de03c0a/openstack-network-exporter/0.log" Oct 14 09:05:20 crc kubenswrapper[5018]: I1014 09:05:20.092748 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_64ed49b9-ade8-4fe4-9c37-ae327de03c0a/ovsdbserver-sb/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.386787 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/util/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.609311 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/util/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.620915 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/pull/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.671737 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/pull/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.770340 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/util/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.810507 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/pull/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.859564 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036a4ttqg_7ba36249-21fb-410f-895a-19cccb4e5fb1/extract/0.log" Oct 14 09:05:33 crc kubenswrapper[5018]: I1014 09:05:33.973916 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-swhnp_a0e4cf91-8413-4a0f-b5ec-06f8becca6a7/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.090045 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-vwpjg_0e01bea2-81b7-4b83-881c-c40f3a8bd306/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.124060 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-swhnp_a0e4cf91-8413-4a0f-b5ec-06f8becca6a7/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.239213 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-vwpjg_0e01bea2-81b7-4b83-881c-c40f3a8bd306/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.341542 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-4fgw9_a81fad02-cd39-487a-ba7d-d29757b26848/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.342930 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-4fgw9_a81fad02-cd39-487a-ba7d-d29757b26848/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.405298 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-gnt2q_d94d1327-34a3-45fe-963c-b80cc8931811/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.584843 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-gnt2q_d94d1327-34a3-45fe-963c-b80cc8931811/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.601144 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-c6q56_284f0e2f-1e97-4201-8313-0c604eb245c7/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.674808 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-c6q56_284f0e2f-1e97-4201-8313-0c604eb245c7/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.787553 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-b9n5t_88f2a077-948e-4245-834d-9f943670b40f/manager/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.795364 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-b9n5t_88f2a077-948e-4245-834d-9f943670b40f/kube-rbac-proxy/0.log" Oct 14 09:05:34 crc kubenswrapper[5018]: I1014 09:05:34.974988 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tfcm7_dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.029878 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-jrnfz_a3965916-7dc3-4bfc-ae94-c57ef9b4b43c/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.158508 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tfcm7_dd9ddc7a-1d4a-479e-a1d5-0aecfa8c32ff/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.202671 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-jrnfz_a3965916-7dc3-4bfc-ae94-c57ef9b4b43c/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.268053 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-wrxs4_4afa0a87-9ea4-4d10-857d-8c046132dbd3/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.393103 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-v2ppk_3195f44a-f801-4f06-ae14-d83a680343a7/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.460867 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-wrxs4_4afa0a87-9ea4-4d10-857d-8c046132dbd3/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.482286 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-v2ppk_3195f44a-f801-4f06-ae14-d83a680343a7/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.607210 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-92pt5_6bf6a59f-81bd-4003-a387-32c703648d21/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.664447 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-92pt5_6bf6a59f-81bd-4003-a387-32c703648d21/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.777895 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-rfcrs_3c93e5a0-bd99-4348-8562-6584ca640fc8/kube-rbac-proxy/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.842607 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-rfcrs_3c93e5a0-bd99-4348-8562-6584ca640fc8/manager/0.log" Oct 14 09:05:35 crc kubenswrapper[5018]: I1014 09:05:35.902871 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-dhc98_15fc314b-1904-488c-829c-9c5c0274ff20/kube-rbac-proxy/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.088739 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-mc6cb_5fea696b-3ec3-47c7-bc33-212456367942/manager/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.127452 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-mc6cb_5fea696b-3ec3-47c7-bc33-212456367942/kube-rbac-proxy/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.129590 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-dhc98_15fc314b-1904-488c-829c-9c5c0274ff20/manager/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.280279 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d448487bng8_fb295ac9-43c8-4482-b557-9ac8aff33171/kube-rbac-proxy/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.300417 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d448487bng8_fb295ac9-43c8-4482-b557-9ac8aff33171/manager/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.468138 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-lxsvb_087226e7-48a0-447d-9a17-59492dac054b/kube-rbac-proxy/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.517017 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64895cd698-5vxbw_50db8b96-a5ef-46d9-ba42-4f41d61ff536/kube-rbac-proxy/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.766750 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64895cd698-5vxbw_50db8b96-a5ef-46d9-ba42-4f41d61ff536/operator/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.913504 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vlqj5_72081dcc-e05e-4d65-bc3b-cf6a398a66d2/registry-server/0.log" Oct 14 09:05:36 crc kubenswrapper[5018]: I1014 09:05:36.983768 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-wltkk_df3020fc-9cba-409f-8355-c76ddfb47db2/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.044713 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-wltkk_df3020fc-9cba-409f-8355-c76ddfb47db2/manager/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.116441 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-vx59l_e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.253973 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-vx59l_e31b129a-e9ad-4f2a-a78c-04f1aeb8ac8f/manager/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.284102 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-nxt7x_1c6a82d2-fe9c-4915-8942-865c875efb80/operator/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.489895 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-s48sn_fdb571a6-0d71-4cd9-885d-798b4489b816/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.542067 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-pgbzn_dba339e1-27ec-4b88-aa67-b1d72aced19a/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.544088 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-s48sn_fdb571a6-0d71-4cd9-885d-798b4489b816/manager/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.719272 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-nlxhj_a5e07832-8614-4f4a-9d58-9fad29c4c425/manager/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.776003 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-nlxhj_a5e07832-8614-4f4a-9d58-9fad29c4c425/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.938212 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-mqvhc_a7f848f3-9deb-4528-84b4-13f28c5900e9/kube-rbac-proxy/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.948904 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-pgbzn_dba339e1-27ec-4b88-aa67-b1d72aced19a/manager/0.log" Oct 14 09:05:37 crc kubenswrapper[5018]: I1014 09:05:37.962795 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-mqvhc_a7f848f3-9deb-4528-84b4-13f28c5900e9/manager/0.log" Oct 14 09:05:38 crc kubenswrapper[5018]: I1014 09:05:38.218364 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-lxsvb_087226e7-48a0-447d-9a17-59492dac054b/manager/0.log" Oct 14 09:05:54 crc kubenswrapper[5018]: I1014 09:05:54.158824 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j2fgv_22b208b1-18a2-4e13-af40-f337838b8218/control-plane-machine-set-operator/0.log" Oct 14 09:05:54 crc kubenswrapper[5018]: I1014 09:05:54.313164 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dp82w_7cd77e7d-730d-4259-b784-4a7b8add26e0/machine-api-operator/0.log" Oct 14 09:05:54 crc kubenswrapper[5018]: I1014 09:05:54.346891 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dp82w_7cd77e7d-730d-4259-b784-4a7b8add26e0/kube-rbac-proxy/0.log" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.087904 5018 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:06 crc kubenswrapper[5018]: E1014 09:06:06.088880 5018 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab39f288-65d7-4f16-86f0-c1d834239898" containerName="container-00" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.088899 5018 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab39f288-65d7-4f16-86f0-c1d834239898" containerName="container-00" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.089109 5018 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab39f288-65d7-4f16-86f0-c1d834239898" containerName="container-00" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.090509 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.102889 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.263391 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.263783 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w7s2\" (UniqueName: \"kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.263923 5018 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.365661 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.365804 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.365841 5018 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w7s2\" (UniqueName: \"kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.366447 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.366467 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.383652 5018 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w7s2\" (UniqueName: \"kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2\") pod \"community-operators-fm79w\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.410538 5018 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.674769 5018 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:06 crc kubenswrapper[5018]: I1014 09:06:06.695117 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerStarted","Data":"3f07a01ee3c4c7f1a54aaf59c5e9fbab50605b737a4bc91c5b8f456e858b7696"} Oct 14 09:06:07 crc kubenswrapper[5018]: I1014 09:06:07.372514 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-czf65_3212416e-b631-4294-83fc-75b350569b44/cert-manager-controller/0.log" Oct 14 09:06:07 crc kubenswrapper[5018]: I1014 09:06:07.456667 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-29g6d_07f3eb84-992a-499b-8916-a349d9defeec/cert-manager-cainjector/0.log" Oct 14 09:06:07 crc kubenswrapper[5018]: I1014 09:06:07.511161 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-sb596_e7159c40-ec11-43d3-84b4-1b3696caf977/cert-manager-webhook/0.log" Oct 14 09:06:07 crc kubenswrapper[5018]: I1014 09:06:07.705474 5018 generic.go:334] "Generic (PLEG): container finished" podID="56da0c97-c396-4ccc-bc6f-152d945204c1" containerID="4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c" exitCode=0 Oct 14 09:06:07 crc kubenswrapper[5018]: I1014 09:06:07.705540 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerDied","Data":"4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c"} Oct 14 09:06:08 crc kubenswrapper[5018]: I1014 09:06:08.713742 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerStarted","Data":"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe"} Oct 14 09:06:09 crc kubenswrapper[5018]: I1014 09:06:09.723464 5018 generic.go:334] "Generic (PLEG): container finished" podID="56da0c97-c396-4ccc-bc6f-152d945204c1" containerID="5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe" exitCode=0 Oct 14 09:06:09 crc kubenswrapper[5018]: I1014 09:06:09.723817 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerDied","Data":"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe"} Oct 14 09:06:10 crc kubenswrapper[5018]: I1014 09:06:10.750862 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerStarted","Data":"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce"} Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.411580 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.412228 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.457785 5018 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.490764 5018 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fm79w" podStartSLOduration=7.927545988 podStartE2EDuration="10.490742148s" podCreationTimestamp="2025-10-14 09:06:06 +0000 UTC" firstStartedPulling="2025-10-14 09:06:07.707238664 +0000 UTC m=+8184.291285281" lastFinishedPulling="2025-10-14 09:06:10.270434814 +0000 UTC m=+8186.854481441" observedRunningTime="2025-10-14 09:06:10.781455116 +0000 UTC m=+8187.365501763" watchObservedRunningTime="2025-10-14 09:06:16.490742148 +0000 UTC m=+8193.074788785" Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.888759 5018 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:16 crc kubenswrapper[5018]: I1014 09:06:16.963476 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:18 crc kubenswrapper[5018]: I1014 09:06:18.704103 5018 scope.go:117] "RemoveContainer" containerID="3ff4e53d026046161fc0ba30442e5f12b23e128b7dfd5eb8afd92d80365c87b0" Oct 14 09:06:18 crc kubenswrapper[5018]: I1014 09:06:18.728385 5018 scope.go:117] "RemoveContainer" containerID="a37980b420c8b815e40b9a904710947c3fc4d3660e112850bdb5cb237e5eafaf" Oct 14 09:06:18 crc kubenswrapper[5018]: I1014 09:06:18.826594 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fm79w" podUID="56da0c97-c396-4ccc-bc6f-152d945204c1" containerName="registry-server" containerID="cri-o://05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce" gracePeriod=2 Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.273058 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.393695 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities\") pod \"56da0c97-c396-4ccc-bc6f-152d945204c1\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.393799 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w7s2\" (UniqueName: \"kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2\") pod \"56da0c97-c396-4ccc-bc6f-152d945204c1\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.393849 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content\") pod \"56da0c97-c396-4ccc-bc6f-152d945204c1\" (UID: \"56da0c97-c396-4ccc-bc6f-152d945204c1\") " Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.395519 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities" (OuterVolumeSpecName: "utilities") pod "56da0c97-c396-4ccc-bc6f-152d945204c1" (UID: "56da0c97-c396-4ccc-bc6f-152d945204c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.401080 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2" (OuterVolumeSpecName: "kube-api-access-6w7s2") pod "56da0c97-c396-4ccc-bc6f-152d945204c1" (UID: "56da0c97-c396-4ccc-bc6f-152d945204c1"). InnerVolumeSpecName "kube-api-access-6w7s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.461734 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56da0c97-c396-4ccc-bc6f-152d945204c1" (UID: "56da0c97-c396-4ccc-bc6f-152d945204c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.495509 5018 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.495546 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w7s2\" (UniqueName: \"kubernetes.io/projected/56da0c97-c396-4ccc-bc6f-152d945204c1-kube-api-access-6w7s2\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.495556 5018 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56da0c97-c396-4ccc-bc6f-152d945204c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.836919 5018 generic.go:334] "Generic (PLEG): container finished" podID="56da0c97-c396-4ccc-bc6f-152d945204c1" containerID="05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce" exitCode=0 Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.836971 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerDied","Data":"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce"} Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.836999 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fm79w" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.837022 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fm79w" event={"ID":"56da0c97-c396-4ccc-bc6f-152d945204c1","Type":"ContainerDied","Data":"3f07a01ee3c4c7f1a54aaf59c5e9fbab50605b737a4bc91c5b8f456e858b7696"} Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.837044 5018 scope.go:117] "RemoveContainer" containerID="05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.857830 5018 scope.go:117] "RemoveContainer" containerID="5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.876124 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.882052 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fm79w"] Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.910533 5018 scope.go:117] "RemoveContainer" containerID="4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.932088 5018 scope.go:117] "RemoveContainer" containerID="05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce" Oct 14 09:06:19 crc kubenswrapper[5018]: E1014 09:06:19.932800 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce\": container with ID starting with 05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce not found: ID does not exist" containerID="05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.932842 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce"} err="failed to get container status \"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce\": rpc error: code = NotFound desc = could not find container \"05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce\": container with ID starting with 05cb7ae3c486dba6eaa8fc4d50dda2fefeb6140b6af559460a577b285f2532ce not found: ID does not exist" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.932870 5018 scope.go:117] "RemoveContainer" containerID="5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe" Oct 14 09:06:19 crc kubenswrapper[5018]: E1014 09:06:19.933246 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe\": container with ID starting with 5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe not found: ID does not exist" containerID="5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.933277 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe"} err="failed to get container status \"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe\": rpc error: code = NotFound desc = could not find container \"5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe\": container with ID starting with 5760c4bdcb6bd0c315f2d3d74df2c2b08472f4773637a1b541fd897bb1e52afe not found: ID does not exist" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.933299 5018 scope.go:117] "RemoveContainer" containerID="4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c" Oct 14 09:06:19 crc kubenswrapper[5018]: E1014 09:06:19.933555 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c\": container with ID starting with 4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c not found: ID does not exist" containerID="4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c" Oct 14 09:06:19 crc kubenswrapper[5018]: I1014 09:06:19.953085 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c"} err="failed to get container status \"4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c\": rpc error: code = NotFound desc = could not find container \"4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c\": container with ID starting with 4dddc231ae78fc5f3220e03d9d54364887847ad34dad960400b87795dfb3ec9c not found: ID does not exist" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.504483 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-gbl7v_b6eb676d-62fd-4035-8cfd-4617d1cb7b35/nmstate-console-plugin/0.log" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.613475 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56da0c97-c396-4ccc-bc6f-152d945204c1" path="/var/lib/kubelet/pods/56da0c97-c396-4ccc-bc6f-152d945204c1/volumes" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.642322 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-556sb_c32afd97-261a-4e41-bda2-ae5de91f5d6f/nmstate-handler/0.log" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.679442 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-lszjp_c718fdd6-a84c-4842-aa5d-f6ccc0c8babd/kube-rbac-proxy/0.log" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.728378 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-lszjp_c718fdd6-a84c-4842-aa5d-f6ccc0c8babd/nmstate-metrics/0.log" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.876452 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-8j4wk_b06cab00-8e52-4c9a-b263-e54c98b12aa5/nmstate-operator/0.log" Oct 14 09:06:20 crc kubenswrapper[5018]: I1014 09:06:20.917717 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-rcf8x_74782f46-6533-44c4-b5bf-bcdd43231af6/nmstate-webhook/0.log" Oct 14 09:06:32 crc kubenswrapper[5018]: I1014 09:06:32.463582 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:06:32 crc kubenswrapper[5018]: I1014 09:06:32.464748 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:06:35 crc kubenswrapper[5018]: I1014 09:06:35.760659 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2xckb_b7499690-3aab-4d77-aa51-9381ab15492a/kube-rbac-proxy/0.log" Oct 14 09:06:35 crc kubenswrapper[5018]: I1014 09:06:35.982837 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-frr-files/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.212985 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-metrics/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.241362 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-frr-files/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.254951 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-reloader/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.404759 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2xckb_b7499690-3aab-4d77-aa51-9381ab15492a/controller/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.421668 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-reloader/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.581793 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-reloader/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.590244 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-metrics/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.592592 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-frr-files/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.620431 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-metrics/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.778952 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-metrics/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.807307 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-reloader/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.809350 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/cp-frr-files/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.825373 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/controller/0.log" Oct 14 09:06:36 crc kubenswrapper[5018]: I1014 09:06:36.989965 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/frr-metrics/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.019552 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/kube-rbac-proxy/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.055798 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/kube-rbac-proxy-frr/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.245674 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/reloader/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.274181 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-84kzf_d3889b7c-1f1b-4ea6-aca2-51d2fa7a658d/frr-k8s-webhook-server/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.477465 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f688cc4bd-dzcxj_5fb208e9-58f7-43be-81d4-ecfa72a34dba/manager/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.706116 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6784749796-zrctk_62f6d4e3-ca4f-44c8-9206-eda31a949201/webhook-server/0.log" Oct 14 09:06:37 crc kubenswrapper[5018]: I1014 09:06:37.727538 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k9tt_05f5473d-ed86-4967-9785-062bc507742f/kube-rbac-proxy/0.log" Oct 14 09:06:38 crc kubenswrapper[5018]: I1014 09:06:38.614494 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k9tt_05f5473d-ed86-4967-9785-062bc507742f/speaker/0.log" Oct 14 09:06:39 crc kubenswrapper[5018]: I1014 09:06:39.882816 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sv8bv_db040ce3-0234-42ce-a316-9c05abc7dda5/frr/0.log" Oct 14 09:06:51 crc kubenswrapper[5018]: I1014 09:06:51.678281 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/util/0.log" Oct 14 09:06:51 crc kubenswrapper[5018]: I1014 09:06:51.827600 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/pull/0.log" Oct 14 09:06:51 crc kubenswrapper[5018]: I1014 09:06:51.883294 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/util/0.log" Oct 14 09:06:51 crc kubenswrapper[5018]: I1014 09:06:51.886134 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.091962 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/extract/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.093977 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.119760 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb694hjtw_363b573a-a6f7-4633-8eb5-4d0f804f546d/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.253337 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.411679 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.449527 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.462052 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.624710 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.644567 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/extract/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.656154 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fpphb_282c68db-308a-4370-99fd-c20f6a5ff81d/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.801568 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.964806 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/util/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.974808 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/pull/0.log" Oct 14 09:06:52 crc kubenswrapper[5018]: I1014 09:06:52.999455 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/pull/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.187577 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/extract/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.191144 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/util/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.256931 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dxtlp4_ed44b30f-2a96-4e81-9516-03fd51a08883/pull/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.368380 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-utilities/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.537839 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-content/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.570495 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-content/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.579892 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-utilities/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.756491 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-content/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.796937 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/extract-utilities/0.log" Oct 14 09:06:53 crc kubenswrapper[5018]: I1014 09:06:53.953583 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-utilities/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.104990 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-content/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.113903 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-utilities/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.168926 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-content/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.377533 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-content/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.420901 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/extract-utilities/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.695176 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/util/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.851223 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/pull/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.884878 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/util/0.log" Oct 14 09:06:54 crc kubenswrapper[5018]: I1014 09:06:54.906763 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/pull/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.089387 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5g2cw_77fd3585-d309-4ec2-892a-c255f69032b3/registry-server/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.117508 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/util/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.131253 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/pull/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.187851 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cm24xd_f742418e-85fe-4763-9fc5-09795c16569b/extract/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.313668 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nxpmh_d6278096-b891-4ca6-a1e1-37b2c6a4277c/registry-server/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.369253 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bh2kb_efe769b7-56d1-4866-bde7-9fc17c66f8cf/marketplace-operator/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.417635 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-utilities/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.544359 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-utilities/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.550276 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-content/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.567036 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-content/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.701387 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-content/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.709919 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/extract-utilities/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.830254 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-utilities/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.952872 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-csprd_171f6e90-d807-4e7d-8e5e-20951b92fb7a/registry-server/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.957181 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-content/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.970920 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-content/0.log" Oct 14 09:06:55 crc kubenswrapper[5018]: I1014 09:06:55.984186 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-utilities/0.log" Oct 14 09:06:56 crc kubenswrapper[5018]: I1014 09:06:56.121691 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-content/0.log" Oct 14 09:06:56 crc kubenswrapper[5018]: I1014 09:06:56.142429 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/extract-utilities/0.log" Oct 14 09:06:56 crc kubenswrapper[5018]: I1014 09:06:56.930136 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hswfh_d5c73ed3-ccf0-48fb-9edf-93748b24edf7/registry-server/0.log" Oct 14 09:07:02 crc kubenswrapper[5018]: I1014 09:07:02.463559 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:07:02 crc kubenswrapper[5018]: I1014 09:07:02.464512 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:07:09 crc kubenswrapper[5018]: I1014 09:07:09.667065 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-skzn7_6922faee-23ec-45f2-9542-5d2e48ccf319/prometheus-operator/0.log" Oct 14 09:07:09 crc kubenswrapper[5018]: I1014 09:07:09.813915 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6bf767b955-rptmv_1ebe5d8f-9403-41da-8c97-ed60af150539/prometheus-operator-admission-webhook/0.log" Oct 14 09:07:09 crc kubenswrapper[5018]: I1014 09:07:09.827978 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6bf767b955-zst8q_de42af88-f42f-4a3a-8a7a-77dec0807d7a/prometheus-operator-admission-webhook/0.log" Oct 14 09:07:10 crc kubenswrapper[5018]: I1014 09:07:10.010761 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-dwfhs_41b12c8d-37f0-4fe5-91a2-3dc8585f9f9d/operator/0.log" Oct 14 09:07:10 crc kubenswrapper[5018]: I1014 09:07:10.059346 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-vs6vr_b1d0a181-0c76-4535-a5e0-ffd8bff1383e/perses-operator/0.log" Oct 14 09:07:18 crc kubenswrapper[5018]: I1014 09:07:18.772725 5018 scope.go:117] "RemoveContainer" containerID="c3017755a0e2f72721deb1e16c3269c756b7cd3d3374013490a920e104ae5d49" Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.476845 5018 kuberuntime_container.go:700] "PreStop hook not completed in grace period" pod="openstack/ovsdbserver-sb-1" podUID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerName="ovsdbserver-sb" containerID="cri-o://6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8" gracePeriod=300 Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.477211 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-1" podUID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerName="ovsdbserver-sb" containerID="cri-o://6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8" gracePeriod=2 Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.490864 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > execCommand=["/usr/local/bin/container-scripts/cleanup.sh"] containerName="ovsdbserver-sb" pod="openstack/ovsdbserver-sb-1" message=< Oct 14 09:07:20 crc kubenswrapper[5018]: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.491172 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > pod="openstack/ovsdbserver-sb-1" podUID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerName="ovsdbserver-sb" containerID="cri-o://6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8" Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.541479 5018 kuberuntime_container.go:700] "PreStop hook not completed in grace period" pod="openstack/ovsdbserver-sb-2" podUID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerName="ovsdbserver-sb" containerID="cri-o://6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9" gracePeriod=300 Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.541559 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-2" podUID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerName="ovsdbserver-sb" containerID="cri-o://6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9" gracePeriod=2 Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.552971 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > execCommand=["/usr/local/bin/container-scripts/cleanup.sh"] containerName="ovsdbserver-sb" pod="openstack/ovsdbserver-sb-2" message=< Oct 14 09:07:20 crc kubenswrapper[5018]: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.553568 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=sb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnsb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ sb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-sb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\s\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnsb_db.ctl cluster/leave OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnsb_db.ctl cluster/status OVN_Southbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > pod="openstack/ovsdbserver-sb-2" podUID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerName="ovsdbserver-sb" containerID="cri-o://6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9" Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.797804 5018 kuberuntime_container.go:700] "PreStop hook not completed in grace period" pod="openstack/ovsdbserver-nb-1" podUID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerName="ovsdbserver-nb" containerID="cri-o://6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a" gracePeriod=300 Oct 14 09:07:20 crc kubenswrapper[5018]: I1014 09:07:20.798438 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-1" podUID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerName="ovsdbserver-nb" containerID="cri-o://6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a" gracePeriod=2 Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.813512 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > execCommand=["/usr/local/bin/container-scripts/cleanup.sh"] containerName="ovsdbserver-nb" pod="openstack/ovsdbserver-nb-1" message=< Oct 14 09:07:20 crc kubenswrapper[5018]: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: > Oct 14 09:07:20 crc kubenswrapper[5018]: E1014 09:07:20.815149 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 09:07:20 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:20 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:20 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:20 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:20 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:20 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-1 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:20 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:20 crc kubenswrapper[5018]: + true Oct 14 09:07:20 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:20 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:20 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:20 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:20 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:20 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: > pod="openstack/ovsdbserver-nb-1" podUID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerName="ovsdbserver-nb" containerID="cri-o://6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:20.869248 5018 kuberuntime_container.go:700] "PreStop hook not completed in grace period" pod="openstack/ovsdbserver-nb-2" podUID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerName="ovsdbserver-nb" containerID="cri-o://2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad" gracePeriod=300 Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:20.869319 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-2" podUID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerName="ovsdbserver-nb" containerID="cri-o://2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad" gracePeriod=2 Oct 14 09:07:21 crc kubenswrapper[5018]: E1014 09:07:20.885697 5018 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 09:07:21 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:21 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:21 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:21 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:21 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: > execCommand=["/usr/local/bin/container-scripts/cleanup.sh"] containerName="ovsdbserver-nb" pod="openstack/ovsdbserver-nb-2" message=< Oct 14 09:07:21 crc kubenswrapper[5018]: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:21 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:21 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:21 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:21 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: > Oct 14 09:07:21 crc kubenswrapper[5018]: E1014 09:07:20.885999 5018 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 09:07:21 crc kubenswrapper[5018]: command '/usr/local/bin/container-scripts/cleanup.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/cleanup.sh Oct 14 09:07:21 crc kubenswrapper[5018]: + source /usr/local/bin/container-scripts/functions Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_TYPE=nb Oct 14 09:07:21 crc kubenswrapper[5018]: ++ DB_FILE=/etc/ovn/ovnnb_db.db Oct 14 09:07:21 crc kubenswrapper[5018]: + DB_NAME=OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ nb == \s\b ]] Oct 14 09:07:21 crc kubenswrapper[5018]: ++ hostname Oct 14 09:07:21 crc kubenswrapper[5018]: + [[ ovsdbserver-nb-2 != \o\v\s\d\b\s\e\r\v\e\r\-\n\b\-\0 ]] Oct 14 09:07:21 crc kubenswrapper[5018]: + ovs-appctl -t /tmp/ovnnb_db.ctl cluster/leave OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: + true Oct 14 09:07:21 crc kubenswrapper[5018]: ++ ovs-appctl -t /tmp/ovnnb_db.ctl cluster/status OVN_Northbound Oct 14 09:07:21 crc kubenswrapper[5018]: ++ grep Status: Oct 14 09:07:21 crc kubenswrapper[5018]: ++ awk -e '{print $2}' Oct 14 09:07:21 crc kubenswrapper[5018]: + STATUS=leaving Oct 14 09:07:21 crc kubenswrapper[5018]: + '[' -z leaving -o xleaving = 'xleft cluster' ']' Oct 14 09:07:21 crc kubenswrapper[5018]: + sleep 1 Oct 14 09:07:21 crc kubenswrapper[5018]: > pod="openstack/ovsdbserver-nb-2" podUID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerName="ovsdbserver-nb" containerID="cri-o://2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.393152 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_495a63bc-82b8-4a2b-83ab-5aed23f6f493/ovsdbserver-nb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.393213 5018 generic.go:334] "Generic (PLEG): container finished" podID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" containerID="2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad" exitCode=143 Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.393279 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerDied","Data":"2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad"} Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.396357 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_64ed49b9-ade8-4fe4-9c37-ae327de03c0a/ovsdbserver-sb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.396462 5018 generic.go:334] "Generic (PLEG): container finished" podID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" containerID="6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9" exitCode=143 Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.396523 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerDied","Data":"6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9"} Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.399956 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bcdada18-a44b-4d39-b4f6-2bc9fef4530c/ovsdbserver-nb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.399998 5018 generic.go:334] "Generic (PLEG): container finished" podID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" containerID="6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a" exitCode=143 Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.400050 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerDied","Data":"6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a"} Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.404375 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/ovsdbserver-sb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.404468 5018 generic.go:334] "Generic (PLEG): container finished" podID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" containerID="6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8" exitCode=143 Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.404537 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerDied","Data":"6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8"} Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.573966 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_495a63bc-82b8-4a2b-83ab-5aed23f6f493/ovsdbserver-nb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.574027 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.577991 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_64ed49b9-ade8-4fe4-9c37-ae327de03c0a/ovsdbserver-sb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.578031 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.618072 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/ovsdbserver-sb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.618130 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.623045 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bcdada18-a44b-4d39-b4f6-2bc9fef4530c/ovsdbserver-nb/0.log" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.623089 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.699372 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700265 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config" (OuterVolumeSpecName: "config") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700591 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700666 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700699 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fqf2\" (UniqueName: \"kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700734 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700778 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700817 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.700843 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.701193 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config" (OuterVolumeSpecName: "config") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.701821 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.702141 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts" (OuterVolumeSpecName: "scripts") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.705024 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706089 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706131 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706148 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706194 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706222 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706244 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n977\" (UniqueName: \"kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706287 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706309 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706343 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706369 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706394 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706418 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cdvt\" (UniqueName: \"kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706440 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706465 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config\") pod \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\" (UID: \"bcdada18-a44b-4d39-b4f6-2bc9fef4530c\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706712 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config" (OuterVolumeSpecName: "config") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.706743 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707040 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707745 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707782 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707800 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707817 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707839 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707874 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts\") pod \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\" (UID: \"495a63bc-82b8-4a2b-83ab-5aed23f6f493\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707893 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf5lp\" (UniqueName: \"kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp\") pod \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\" (UID: \"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707917 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.707934 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts\") pod \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\" (UID: \"64ed49b9-ade8-4fe4-9c37-ae327de03c0a\") " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708465 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708484 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708492 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708501 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708512 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708520 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.708528 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.709018 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.709092 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts" (OuterVolumeSpecName: "scripts") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.709602 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts" (OuterVolumeSpecName: "scripts") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.709787 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts" (OuterVolumeSpecName: "scripts") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.710449 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config" (OuterVolumeSpecName: "config") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.710706 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt" (OuterVolumeSpecName: "kube-api-access-4cdvt") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "kube-api-access-4cdvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.710849 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2" (OuterVolumeSpecName: "kube-api-access-9fqf2") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "kube-api-access-9fqf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.715531 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp" (OuterVolumeSpecName: "kube-api-access-vf5lp") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "kube-api-access-vf5lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.715862 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "pvc-7442a043-95cd-4b70-9823-eab754fdabb4". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.718920 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977" (OuterVolumeSpecName: "kube-api-access-8n977") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "kube-api-access-8n977". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.738951 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.738974 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "pvc-f1800c50-fd1f-472b-a205-a6e34313b118". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.747855 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.752796 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.755983 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.761914 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.775593 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.779317 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.780520 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.783544 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.789681 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "64ed49b9-ade8-4fe4-9c37-ae327de03c0a" (UID: "64ed49b9-ade8-4fe4-9c37-ae327de03c0a"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.793882 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.800289 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" (UID: "83a2d465-d49e-4f9d-a1c9-d5f674ae30e9"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.805158 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "495a63bc-82b8-4a2b-83ab-5aed23f6f493" (UID: "495a63bc-82b8-4a2b-83ab-5aed23f6f493"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.807171 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bcdada18-a44b-4d39-b4f6-2bc9fef4530c" (UID: "bcdada18-a44b-4d39-b4f6-2bc9fef4530c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810215 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") on node \"crc\" " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810241 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") on node \"crc\" " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810255 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810268 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810278 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n977\" (UniqueName: \"kubernetes.io/projected/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-kube-api-access-8n977\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810288 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810298 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810306 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810314 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810322 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810331 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cdvt\" (UniqueName: \"kubernetes.io/projected/495a63bc-82b8-4a2b-83ab-5aed23f6f493-kube-api-access-4cdvt\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810338 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495a63bc-82b8-4a2b-83ab-5aed23f6f493-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810347 5018 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810401 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") on node \"crc\" " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810411 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810421 5018 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810429 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810437 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810446 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/495a63bc-82b8-4a2b-83ab-5aed23f6f493-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810454 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf5lp\" (UniqueName: \"kubernetes.io/projected/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9-kube-api-access-vf5lp\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810462 5018 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810470 5018 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64ed49b9-ade8-4fe4-9c37-ae327de03c0a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810484 5018 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") on node \"crc\" " Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810493 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fqf2\" (UniqueName: \"kubernetes.io/projected/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-kube-api-access-9fqf2\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.810502 5018 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcdada18-a44b-4d39-b4f6-2bc9fef4530c-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.841722 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.841816 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.841874 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7442a043-95cd-4b70-9823-eab754fdabb4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4") on node "crc" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.842006 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53") on node "crc" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.847926 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.848119 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f1800c50-fd1f-472b-a205-a6e34313b118" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118") on node "crc" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.848119 5018 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.848248 5018 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a") on node "crc" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.911870 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae4b0879-a971-4f91-9b7a-b3dd48b7cb53\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.911909 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-7442a043-95cd-4b70-9823-eab754fdabb4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7442a043-95cd-4b70-9823-eab754fdabb4\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.911919 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-f1800c50-fd1f-472b-a205-a6e34313b118\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1800c50-fd1f-472b-a205-a6e34313b118\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:21 crc kubenswrapper[5018]: I1014 09:07:21.911928 5018 reconciler_common.go:293] "Volume detached for volume \"pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ef5f72-4daf-4d2b-b2b5-c63a433a742a\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.419443 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_64ed49b9-ade8-4fe4-9c37-ae327de03c0a/ovsdbserver-sb/0.log" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.419592 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"64ed49b9-ade8-4fe4-9c37-ae327de03c0a","Type":"ContainerDied","Data":"d77a1d281e4fe33ee612e40d06af4451ebeb1d0d1001be5ccb6285dd06990598"} Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.419682 5018 scope.go:117] "RemoveContainer" containerID="19de2087aaf37731d242d90edcb3f4e4a539c73a836422704d5ea34ead7356dc" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.420776 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.424689 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bcdada18-a44b-4d39-b4f6-2bc9fef4530c/ovsdbserver-nb/0.log" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.424890 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.424972 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bcdada18-a44b-4d39-b4f6-2bc9fef4530c","Type":"ContainerDied","Data":"f127d2b035406ce1d2e06959048355959f5819539681b83eac1c3a7ba3864e9b"} Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.429811 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/ovsdbserver-sb/0.log" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.430191 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.431565 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"83a2d465-d49e-4f9d-a1c9-d5f674ae30e9","Type":"ContainerDied","Data":"5c651b3c84b92683765b1a9ea0862b522bf4b38c23c9ae7ee3e86fdefce2de3e"} Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.437311 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_495a63bc-82b8-4a2b-83ab-5aed23f6f493/ovsdbserver-nb/0.log" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.437394 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"495a63bc-82b8-4a2b-83ab-5aed23f6f493","Type":"ContainerDied","Data":"64b4317a8e03b1fdb07ea4956329cfde29ff6af5f99d87eb57b15f3b32f4974f"} Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.437532 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.466167 5018 scope.go:117] "RemoveContainer" containerID="6f4633e47fca5fc55363bb55990c10fe6fce5aa14e1d53685595afeec056f7a9" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.515724 5018 scope.go:117] "RemoveContainer" containerID="f0b29be9a0afeb5e7192333db87e868120fdaf92944e86c96054f44887940f2f" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.521751 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.547451 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.557027 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.558827 5018 scope.go:117] "RemoveContainer" containerID="6be7ba3cf6e1701372f16a4f24c67b8879709a8443c5940ebf0e69c9a600357a" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.563959 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.569813 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.574581 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.577235 5018 scope.go:117] "RemoveContainer" containerID="f5a9062ecbff01d430f8c45450a3ea8dfae4ad161b221c5afda3f75ccde512b0" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.579807 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.585893 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.597771 5018 scope.go:117] "RemoveContainer" containerID="6d500b201dd215d7943a79cf551e2d6d79b9ea9ccbb7d365407caa532f3013a8" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.613285 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495a63bc-82b8-4a2b-83ab-5aed23f6f493" path="/var/lib/kubelet/pods/495a63bc-82b8-4a2b-83ab-5aed23f6f493/volumes" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.613935 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ed49b9-ade8-4fe4-9c37-ae327de03c0a" path="/var/lib/kubelet/pods/64ed49b9-ade8-4fe4-9c37-ae327de03c0a/volumes" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.615081 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a2d465-d49e-4f9d-a1c9-d5f674ae30e9" path="/var/lib/kubelet/pods/83a2d465-d49e-4f9d-a1c9-d5f674ae30e9/volumes" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.615735 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcdada18-a44b-4d39-b4f6-2bc9fef4530c" path="/var/lib/kubelet/pods/bcdada18-a44b-4d39-b4f6-2bc9fef4530c/volumes" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.620970 5018 scope.go:117] "RemoveContainer" containerID="8518cc00808c52898e754e85cacd8523a46b73a34084d59361518ea964572758" Oct 14 09:07:22 crc kubenswrapper[5018]: I1014 09:07:22.644081 5018 scope.go:117] "RemoveContainer" containerID="2d550f050dd0f0f7741d281133adcef91cbe86cf0b5dab50eb607c382eb96bad" Oct 14 09:07:32 crc kubenswrapper[5018]: I1014 09:07:32.463550 5018 patch_prober.go:28] interesting pod/machine-config-daemon-g7t5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:07:32 crc kubenswrapper[5018]: I1014 09:07:32.464364 5018 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:07:32 crc kubenswrapper[5018]: I1014 09:07:32.464466 5018 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" Oct 14 09:07:32 crc kubenswrapper[5018]: I1014 09:07:32.465529 5018 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b98d0e5c81187860cd3ad149820c384a66b7867dc7a4fbc14fec8624d602899f"} pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:07:32 crc kubenswrapper[5018]: I1014 09:07:32.465660 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" podUID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerName="machine-config-daemon" containerID="cri-o://b98d0e5c81187860cd3ad149820c384a66b7867dc7a4fbc14fec8624d602899f" gracePeriod=600 Oct 14 09:07:33 crc kubenswrapper[5018]: I1014 09:07:33.579526 5018 generic.go:334] "Generic (PLEG): container finished" podID="124fa7dd-4f8d-4d66-9223-f6bcc6df7242" containerID="b98d0e5c81187860cd3ad149820c384a66b7867dc7a4fbc14fec8624d602899f" exitCode=0 Oct 14 09:07:33 crc kubenswrapper[5018]: I1014 09:07:33.579842 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerDied","Data":"b98d0e5c81187860cd3ad149820c384a66b7867dc7a4fbc14fec8624d602899f"} Oct 14 09:07:33 crc kubenswrapper[5018]: I1014 09:07:33.580450 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-g7t5z" event={"ID":"124fa7dd-4f8d-4d66-9223-f6bcc6df7242","Type":"ContainerStarted","Data":"4245d1cb7669171ee8460373965db0dc8f956ed00756aaab654a2261efaddf90"} Oct 14 09:07:33 crc kubenswrapper[5018]: I1014 09:07:33.580496 5018 scope.go:117] "RemoveContainer" containerID="60c1dae9e8d8ab8bf40e7729e9b3e2820943726539d63f47353defc5881af5ef" Oct 14 09:08:02 crc kubenswrapper[5018]: I1014 09:08:02.935238 5018 generic.go:334] "Generic (PLEG): container finished" podID="e38af198-4b74-412f-9f28-fe5ea89a4e67" containerID="2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989" exitCode=0 Oct 14 09:08:02 crc kubenswrapper[5018]: I1014 09:08:02.935357 5018 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-945sk/must-gather-vs44p" event={"ID":"e38af198-4b74-412f-9f28-fe5ea89a4e67","Type":"ContainerDied","Data":"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989"} Oct 14 09:08:02 crc kubenswrapper[5018]: I1014 09:08:02.936751 5018 scope.go:117] "RemoveContainer" containerID="2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989" Oct 14 09:08:03 crc kubenswrapper[5018]: I1014 09:08:03.777878 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-945sk_must-gather-vs44p_e38af198-4b74-412f-9f28-fe5ea89a4e67/gather/0.log" Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.467527 5018 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-945sk/must-gather-vs44p"] Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.468363 5018 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-945sk/must-gather-vs44p" podUID="e38af198-4b74-412f-9f28-fe5ea89a4e67" containerName="copy" containerID="cri-o://8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6" gracePeriod=2 Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.480083 5018 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-945sk/must-gather-vs44p"] Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.863037 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-945sk_must-gather-vs44p_e38af198-4b74-412f-9f28-fe5ea89a4e67/copy/0.log" Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.863343 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.970581 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7fqr\" (UniqueName: \"kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr\") pod \"e38af198-4b74-412f-9f28-fe5ea89a4e67\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.970647 5018 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output\") pod \"e38af198-4b74-412f-9f28-fe5ea89a4e67\" (UID: \"e38af198-4b74-412f-9f28-fe5ea89a4e67\") " Oct 14 09:08:12 crc kubenswrapper[5018]: I1014 09:08:12.988792 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr" (OuterVolumeSpecName: "kube-api-access-r7fqr") pod "e38af198-4b74-412f-9f28-fe5ea89a4e67" (UID: "e38af198-4b74-412f-9f28-fe5ea89a4e67"). InnerVolumeSpecName "kube-api-access-r7fqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.041490 5018 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-945sk_must-gather-vs44p_e38af198-4b74-412f-9f28-fe5ea89a4e67/copy/0.log" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.041990 5018 generic.go:334] "Generic (PLEG): container finished" podID="e38af198-4b74-412f-9f28-fe5ea89a4e67" containerID="8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6" exitCode=143 Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.042075 5018 scope.go:117] "RemoveContainer" containerID="8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.042268 5018 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-945sk/must-gather-vs44p" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.063308 5018 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e38af198-4b74-412f-9f28-fe5ea89a4e67" (UID: "e38af198-4b74-412f-9f28-fe5ea89a4e67"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.067420 5018 scope.go:117] "RemoveContainer" containerID="2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.072473 5018 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7fqr\" (UniqueName: \"kubernetes.io/projected/e38af198-4b74-412f-9f28-fe5ea89a4e67-kube-api-access-r7fqr\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.072504 5018 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e38af198-4b74-412f-9f28-fe5ea89a4e67-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.111893 5018 scope.go:117] "RemoveContainer" containerID="8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6" Oct 14 09:08:13 crc kubenswrapper[5018]: E1014 09:08:13.112164 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6\": container with ID starting with 8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6 not found: ID does not exist" containerID="8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.112249 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6"} err="failed to get container status \"8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6\": rpc error: code = NotFound desc = could not find container \"8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6\": container with ID starting with 8a778349ce2a4fd7f680a9072ed62b0e8fa19a4a8a890c7cd1b2891efb5ee4c6 not found: ID does not exist" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.112323 5018 scope.go:117] "RemoveContainer" containerID="2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989" Oct 14 09:08:13 crc kubenswrapper[5018]: E1014 09:08:13.112648 5018 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989\": container with ID starting with 2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989 not found: ID does not exist" containerID="2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989" Oct 14 09:08:13 crc kubenswrapper[5018]: I1014 09:08:13.112743 5018 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989"} err="failed to get container status \"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989\": rpc error: code = NotFound desc = could not find container \"2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989\": container with ID starting with 2c05cc3344437f454338ad0a36f5d5fa634d4f155fbfaead705cca8433c7a989 not found: ID does not exist" Oct 14 09:08:14 crc kubenswrapper[5018]: I1014 09:08:14.617410 5018 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e38af198-4b74-412f-9f28-fe5ea89a4e67" path="/var/lib/kubelet/pods/e38af198-4b74-412f-9f28-fe5ea89a4e67/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073411502024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073411502017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073370253016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073370253015462 5ustar corecore